Welcome to the fourth project of the Machine Learning Engineer Nanodegree! In this notebook, template code has already been provided for you to aid in your analysis of the Smartcab and your implemented learning algorithm. You will not need to modify the included code beyond what is requested. There will be questions that you must answer which relate to the project and the visualizations provided in the notebook. Each section where you will answer a question is preceded by a 'Question X' header. Carefully read each question and provide thorough answers in the following text boxes that begin with 'Answer:'. Your project submission will be evaluated based on your answers to each of the questions and the implementation you provide in agent.py.
Note: Code and Markdown cells can be executed using the Shift + Enter keyboard shortcut. In addition, Markdown cells can be edited by typically double-clicking the cell to enter edit mode.
In this project, you will work towards constructing an optimized Q-Learning driving agent that will navigate a Smartcab through its environment towards a goal. Since the Smartcab is expected to drive passengers from one location to another, the driving agent will be evaluated on two very important metrics: Safety and Reliability. A driving agent that gets the Smartcab to its destination while running red lights or narrowly avoiding accidents would be considered unsafe. Similarly, a driving agent that frequently fails to reach the destination in time would be considered unreliable. Maximizing the driving agent's safety and reliability would ensure that Smartcabs have a permanent place in the transportation industry.
Safety and Reliability are measured using a letter-grade system as follows:
| Grade | Safety | Reliability |
|---|---|---|
| A+ | Agent commits no traffic violations, and always chooses the correct action. |
Agent reaches the destination in time for 100% of trips. |
| A | Agent commits few minor traffic violations, such as failing to move on a green light. |
Agent reaches the destination on time for at least 90% of trips. |
| B | Agent commits frequent minor traffic violations, such as failing to move on a green light. |
Agent reaches the destination on time for at least 80% of trips. |
| C | Agent commits at least one major traffic violation, such as driving through a red light. |
Agent reaches the destination on time for at least 70% of trips. |
| D | Agent causes at least one minor accident, such as turning left on green with oncoming traffic. |
Agent reaches the destination on time for at least 60% of trips. |
| F | Agent causes at least one major accident, such as driving through a red light with cross-traffic. |
Agent fails to reach the destination on time for at least 60% of trips. |
To assist evaluating these important metrics, you will need to load visualization code that will be used later on in the project. Run the code cell below to import this code which is required for your analysis.
# Import the visualization code
import visuals as vs
# Pretty display for notebooks
%matplotlib inline
Before starting to work on implementing your driving agent, it's necessary to first understand the world (environment) which the Smartcab and driving agent work in. One of the major components to building a self-learning agent is understanding the characteristics about the agent, which includes how the agent operates. To begin, simply run the agent.py agent code exactly how it is -- no need to make any additions whatsoever. Let the resulting simulation run for some time to see the various working components. Note that in the visual simulation (if enabled), the white vehicle is the Smartcab.
In a few sentences, describe what you observe during the simulation when running the default agent.py agent code. Some things you could consider:
Hint: From the /smartcab/ top-level directory (where this notebook is located), run the command
'python smartcab/agent.py'
%run smartcab/agent.py
/------------------------- | Training trial 1 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.85) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.21) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.24) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.08) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.45) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.82) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.15) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.51) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.30) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.12) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.72) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.21) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.48) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.89) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.06) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.15) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.89) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.18) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.23) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.70) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.12) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.62) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.52) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.20) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.35) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.27) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.63) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.15) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.74) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 2 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.99) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.27) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.27) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.47) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.83) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.29) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.19) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.33) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.30) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.34) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.09) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.51) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.22) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.59) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.42) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.95) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.08) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.86) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.90) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.66) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.71) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.10) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 3 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.12) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.41) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.43) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.47) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.31) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.59) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.73) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.18) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.32) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.27) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.40) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.32) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.26) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.58) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.44) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.47) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.42) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.03) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.34) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.23) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.01) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.08) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.98) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.82) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 4 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.34) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.23) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.70) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.48) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.81) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.83) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.40) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.95) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.67) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.96) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.91) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.34) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.10) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.06) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.50) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 0.61) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.58) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.50) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 5 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.98) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.85) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.65) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.15) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.85) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.15) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.22) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.21) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.94) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.06) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.91) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.37) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.08) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.21) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.15) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.79) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.46) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.00) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.69) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.56) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.01) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.69) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.19) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.36) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.15) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.64) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded -0.78) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 6 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.47) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.12) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.87) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.43) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.33) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.12) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.35) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.32) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.73) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.23) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.76) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.06) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.51) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.04) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.20) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.71) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.64) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.40) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 7 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.07) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.27) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.32) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.44) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.91) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.37) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.22) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.07) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.78) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.51) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.50) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.59) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.25) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 0.99) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.23) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.66) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.37) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.46) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.50) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 8 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.29) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.62) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.15) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.52) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.89) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.98) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.51) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.60) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.24) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.14) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.38) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.73) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.45) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.63) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.05) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.82) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.94) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.00) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.27) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.57) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.77) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.62) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.50) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.39) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 9 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.97) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.62) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.49) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.18) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.09) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.79) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.91) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.93) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.28) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.09) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.22) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.88) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.85) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.82) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.53) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.64) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.99) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.64) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.64) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 10 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.34) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.94) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.27) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.03) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.13) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.23) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.36) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.19) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.62) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.41) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.56) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.00) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.34) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.86) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 0.54) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.50) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.01) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.12) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.58) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 11 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.67) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.70) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.66) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.81) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.70) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.48) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.33) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.36) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.93) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.22) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.86) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.87) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.37) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.32) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.10) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.15) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.79) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.60) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.00) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.03) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.99) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.92) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.43) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.89) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.12) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.82) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.66) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.26) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.80) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 12 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.00) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.14) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.84) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.32) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.45) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.67) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.04) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.92) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.70) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.29) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.46) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.37) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.69) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.19) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.80) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.19) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.20) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.37) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.02) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.37) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.43) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.24) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.42) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.27) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.28) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.89) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.30) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.58) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.07) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.81) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 13 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.59) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.88) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.59) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.41) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.76) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.59) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.06) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.70) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.71) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.71) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.55) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.70) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.30) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.14) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.54) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.77) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.80) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 14 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.16) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.96) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.75) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.48) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.78) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.75) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.11) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.00) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.08) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.60) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.44) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.35) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.50) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.05) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.85) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.31) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.49) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.45) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.02) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.09) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.55) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.63) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.98) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.03) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.72) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.56) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 15 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.90) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.02) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.79) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.01) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.65) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.55) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.10) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.30) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.53) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.49) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.60) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.05) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.92) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.78) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.68) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.42) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.68) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.87) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.68) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 16 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.93) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.95) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.09) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.09) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.52) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.12) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.66) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.93) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.45) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.96) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.05) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.82) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.30) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.14) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.57) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.93) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.18) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.24) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.20) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.27) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 17 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.09) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.53) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.55) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.27) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.79) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.28) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.05) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.95) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.04) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.51) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.10) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.12) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.93) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.12) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.36) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.88) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.43) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.03) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.19) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.48) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.82) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.63) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded -0.04) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 18 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.50) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.97) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.88) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.77) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.85) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.23) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.73) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.37) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.68) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.22) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.15) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.21) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.60) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.46) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.67) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.48) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.23) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.57) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.37) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.20) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.64) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.76) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.15) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.36) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 19 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.16) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.18) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.34) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.17) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.04) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.44) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.60) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.52) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.41) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.32) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.13) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.47) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.43) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.28) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.44) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.17) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 20 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.00) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.64) 94% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.99) 91% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.39) 89% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.12) 86% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.09) 83% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.83) 80% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.54) 77% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.55) 74% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.62) 71% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.80) 69% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.55) 66% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.66) 63% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.11) 60% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.31) 57% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.37) 54% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.33) 51% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.04) 49% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.94) 46% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.46) 43% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.72) 40% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.04) 37% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.39) 34% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.81) 31% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.17) 29% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.47) 26% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.30) 23% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.60) 20% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.28) 17% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.47) 14% of time remaining to reach destination. /------------------- | Step 30 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.07) 11% of time remaining to reach destination. /------------------- | Step 31 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.56) 9% of time remaining to reach destination. /------------------- | Step 32 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.55) 6% of time remaining to reach destination. /------------------- | Step 33 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15) 3% of time remaining to reach destination. /------------------- | Step 34 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.66) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 1 \------------------------- Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.63) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.20) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.35) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.41) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.23) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.92) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.65) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.98) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.52) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.32) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.07) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.54) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.67) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 0.90) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.15) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded -0.37) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.86) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.19) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.20) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.76) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.14) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.75) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.95) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.54) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. Simulation ended. . .
Answer:
In addition to understanding the world, it is also necessary to understand the code itself that governs how the world, simulation, and so on operate. Attempting to create a driving agent would be difficult without having at least explored the "hidden" devices that make everything work. In the /smartcab/ top-level directory, there are two folders: /logs/ (which will be used later) and /smartcab/. Open the /smartcab/ folder and explore each Python file included, then answer the following question.
agent.py Python file, choose three flags that can be set and explain how they change the simulation.environment.py Python file, what Environment class function is called when an agent performs an action?simulator.py Python file, what is the difference between the 'render_text()' function and the 'render()' function?planner.py Python file, will the 'next_waypoint() function consider the North-South or East-West direction first?Answer:
The first step to creating an optimized Q-Learning driving agent is getting the agent to actually take valid actions. In this case, a valid action is one of None, (do nothing) 'left' (turn left), right' (turn right), or 'forward' (go forward). For your first implementation, navigate to the 'choose_action()' agent function and make the driving agent randomly choose one of these actions. Note that you have access to several class variables that will help you write this functionality, such as 'self.learning' and 'self.valid_actions'. Once implemented, run the agent file and simulation briefly to confirm that your driving agent is taking a random action each time step.
To obtain results from the initial simulation, you will need to adjust following flags:
'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.'log_metrics' - Set this to True to log the simluation results as a .csv file in /logs/.'n_test' - Set this to '10' to perform 10 testing trials.Optionally, you may disable to the visual simulation (which can make the trials go faster) by setting the 'display' flag to False. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!
Once you have successfully completed the initial simulation (there should have been 20 training trials and 10 testing trials), run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded! Run the agent.py file after setting the flags from projects/smartcab folder instead of projects/smartcab/smartcab.
%run smartcab/agent.py
/------------------------- | Training trial 1 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.24) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.46) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.44) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.42) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.40) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.08) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.51) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.74) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.49) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.39) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.15) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.56) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.35) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86) 15% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86) 15% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86) 15% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86) 15% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.91) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.99) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 2 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.96) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.34) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.83) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.82) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.38) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 55% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.67) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.18) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.89) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.85) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.38) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.55) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.11) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.99) 5% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.99) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.17) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 3 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.20) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.14) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.67) 88% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.67) 88% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.67) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.15) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.79) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.03) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.68) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.68) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.68) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.68) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.89) 68% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.89) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.36) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.69) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.15) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.50) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.86) 40% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.86) 40% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.86) 40% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.86) 40% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.86) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.11) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.77) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.28) 24% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.28) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.39) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.60) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.08) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.34) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.95) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 4 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.44) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.03) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.61) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.94) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.26) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.16) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.22) 72% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.22) 72% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.22) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.66) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.51) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.53) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.90) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.81) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.73) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.79) 44% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.79) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.19) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.09) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.62) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.11) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.96) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.64) 16% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.64) 16% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.64) 16% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.64) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.88) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.18) 8% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.18) 8% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.18) 8% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.18) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.30) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.95) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 5 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.84) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.67) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.29) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.38) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.07) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.73) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.43) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.59) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.32) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.48) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.94) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.40) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.43) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.28) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.09) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.06) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.09) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.43) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 6 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.80) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.30) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.84) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.77) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.48) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.44) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.58) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.81) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.79) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.40) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.72) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.15) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.90) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.13) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.15) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.53) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.04) 28% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.04) 28% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.04) 28% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.04) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.44) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.38) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 16% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.54) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.00) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.00) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.00) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.49) 8% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.49) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.37) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.42) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 7 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.47) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.01) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.12) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.05) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.97) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.99) 70% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.99) 70% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.99) 70% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.99) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.16) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.09) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29) 55% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29) 55% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29) 55% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29) 55% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.05) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.79) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.19) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.08) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.55) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.27) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.91) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.07) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.18) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.46) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.42) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 8 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.59) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.22) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.43) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.28) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.16) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.18) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.92) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.72) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.79) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.54) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.31) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.52) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.61) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.66) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.73) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.70) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.39) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.01) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.23) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.76) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.35) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.65) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.03) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 9 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.33) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.12) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.42) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.71) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.48) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.30) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.57) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.36) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.25) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.33) 52% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.33) 52% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.33) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.52) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13) 44% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13) 44% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.89) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.31) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.23) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.32) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.70) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded -0.40) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.36) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.68) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.02) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.27) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.43) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 10 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.01) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.20) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.24) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.04) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.75) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.01) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.71) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.24) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.89) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.04) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.85) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.88) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.68) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.91) 44% of time remaining to reach destination. Trial Completed! Agent reached the destination. /------------------------- | Training trial 11 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.34) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.97) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.19) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.67) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.62) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.20) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.96) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.21) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.07) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.76) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.17) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.65) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 0.75) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.30) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.05) 15% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.05) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.95) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.81) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.49) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 12 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.78) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.33) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.28) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.83) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.59) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.18) 76% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.18) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.03) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.13) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.16) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.89) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.11) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.11) 56% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.11) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.14) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.67) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.76) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.61) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.83) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.29) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.67) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.35) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.67) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.08) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.22) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.13) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.31) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 13 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.08) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.63) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.02) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.68) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.65) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.95) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.74) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.69) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.75) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.70) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.28) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.62) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.24) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.58) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.63) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.24) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.71) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.69) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 14 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.81) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.58) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.46) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.66) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.89) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.25) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.65) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.42) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.06) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.48) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.13) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.51) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.91) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.59) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.39) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.22) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.72) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.72) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.72) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.72) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.85) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.27) 33% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.27) 33% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.27) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.35) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.42) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.61) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.13) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.63) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.46) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded -0.40) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.70) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.03) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.87) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 15 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55) 94% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.89) 91% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.52) 89% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.07) 86% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.51) 83% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.96) 80% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.28) 77% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16) 74% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46) 71% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.78) 69% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.78) 69% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.78) 69% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.26) 66% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.39) 63% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.73) 60% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.73) 60% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.73) 60% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.47) 57% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.75) 54% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.04) 51% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.83) 49% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.34) 46% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10) 43% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.32) 40% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.23) 37% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.00) 34% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.55) 31% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.96) 29% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.96) 29% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.96) 29% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.96) 29% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.96) 29% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.34) 26% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.39) 23% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.81) 20% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.14) 17% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.59) 14% of time remaining to reach destination. /------------------- | Step 30 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.94) 11% of time remaining to reach destination. /------------------- | Step 31 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.52) 9% of time remaining to reach destination. /------------------- | Step 32 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.34) 6% of time remaining to reach destination. /------------------- | Step 33 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.97) 3% of time remaining to reach destination. /------------------- | Step 34 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.93) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 16 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.64) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.77) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.02) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.06) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.82) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.61) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.61) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.91) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.24) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.08) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.18) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.08) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.59) 35% of time remaining to reach destination. Trial Completed! Agent reached the destination. /------------------------- | Training trial 17 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.66) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.39) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.43) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.69) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.45) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.87) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.11) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.58) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.69) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.14) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.98) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.71) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.07) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.90) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.57) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.67) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.06) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.30) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.82) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded -0.49) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.80) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.69) 4% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.69) 4% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.69) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.57) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 18 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.23) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.45) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.70) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.23) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.37) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.85) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.35) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.17) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.25) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.57) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.63) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.07) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.03) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.28) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.35) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded -0.25) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.49) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded -0.57) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.21) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.56) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 19 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.20) 97% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.66) 93% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.84) 90% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.63) 87% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.92) 83% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.06) 80% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.06) 77% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.54) 73% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.73) 70% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.26) 67% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.91) 63% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.68) 60% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.12) 57% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05) 53% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.88) 50% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.11) 47% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.79) 43% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84) 40% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.20) 37% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.04) 33% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.00) 30% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.35) 27% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.43) 23% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.43) 23% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.43) 23% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.04) 20% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.72) 17% of time remaining to reach destination. /------------------- | Step 25 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.13) 13% of time remaining to reach destination. /------------------- | Step 26 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.08) 10% of time remaining to reach destination. /------------------- | Step 27 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.91) 7% of time remaining to reach destination. /------------------- | Step 28 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.54) 3% of time remaining to reach destination. /------------------- | Step 29 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Training trial 20 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.51) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.37) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.61) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.95) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.70) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.05) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.11) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.92) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.88) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.17) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.30) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.68) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.23) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.93) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.27) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.52) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.59) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.25) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.43) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.40) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.62) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.74) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.77) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 1 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.87) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.83) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.97) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.61) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.80) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.05) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.88) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.30) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.22) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.58) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.52) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.20) 35% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.20) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.34) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.13) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 0.60) 10% of time remaining to reach destination. Trial Completed! Agent reached the destination. /------------------------- | Testing trial 2 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.78) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.75) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.51) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.52) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.81) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.34) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.60) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.91) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.23) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.60) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.29) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.33) 52% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.33) 52% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.33) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.66) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.34) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.43) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.17) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.75) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.04) 20% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.04) 20% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.04) 20% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.04) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.27) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.27) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.20) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.53) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.76) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 3 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.43) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.83) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.81) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.10) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.61) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.63) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.25) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.90) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.02) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.15) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.49) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.02) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.72) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.52) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.27) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.14) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.64) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.39) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.58) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.94) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 2.27) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.83) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.60) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 4 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.79) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.05) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.87) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.51) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.63) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.00) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.52) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.61) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.39) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.12) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.05) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.05) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.05) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.05) 40% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.05) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.99) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.60) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.28) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.22) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 15% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 15% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 15% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.20) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.36) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.70) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.83) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 5 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.46) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.30) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.17) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.95) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.51) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.28) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.29) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.49) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.13) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.48) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.88) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.74) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.26) 35% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.26) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.66) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.82) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.60) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.11) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent followed the waypoint forward. (rewarded 1.73) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.62) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.57) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 6 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.54) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.33) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.80) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.01) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.06) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.50) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.29) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.25) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.23) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.71) 60% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.71) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.54) 56% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.54) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85) 52% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85) 52% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85) 52% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.64) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded -0.24) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.52) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.66) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.10) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.37) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded -0.09) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.19) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.27) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.18) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.05) 8% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.05) 8% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.05) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.02) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 7 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.00) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.90) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 0.47) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.09) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.35) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.61) 72% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.61) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.36) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.98) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 1.43) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.57) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.31) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.16) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.44) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.31) 36% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -9.31) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.02) 32% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.02) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.04) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.73) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.37) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.40) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.58) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.25) 8% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.25) 8% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.25) 8% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.25) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.71) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded -0.03) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 8 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 1.17) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.27) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.62) 85% of time remaining to reach destination. !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.62) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.20) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.20) 80% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.20) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.00) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.99) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.99) 70% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.99) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.32) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.32) 65% of time remaining to reach destination. !! Agent state not been updated! Agent drove left instead of forward. (rewarded 1.32) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.42) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 1.58) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 1.17) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.75) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.56) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.73) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.87) 30% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.87) 30% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.87) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.16) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.38) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove left instead of right. (rewarded 0.36) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.08) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.41) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.27) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 9 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.37) 95% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.20) 90% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94) 85% of time remaining to reach destination. !! Agent state not been updated! Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94) 85% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 1.81) 80% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 1.66) 75% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.13) 70% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.38) 65% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.98) 60% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.24) 55% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -10.90) 50% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.93) 45% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.86) 40% of time remaining to reach destination. !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.86) 40% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.95) 35% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.15) 30% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded -0.17) 25% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.72) 20% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent attempted driving forward through a red light. (rewarded -10.46) 15% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.17) 10% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.27) 5% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.75) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. /------------------------- | Testing trial 10 \------------------------- Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. Simulating trial. . . Agent not set to learn. /------------------- | Step 0 Results \------------------- !! Agent state not been updated! Agent followed the waypoint right. (rewarded 2.93) 96% of time remaining to reach destination. /------------------- | Step 1 Results \------------------- !! Agent state not been updated! Agent drove right instead of forward. (rewarded 0.54) 92% of time remaining to reach destination. /------------------- | Step 2 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 1.54) 88% of time remaining to reach destination. /------------------- | Step 3 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.19) 84% of time remaining to reach destination. /------------------- | Step 4 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.17) 80% of time remaining to reach destination. /------------------- | Step 5 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.24) 76% of time remaining to reach destination. /------------------- | Step 6 Results \------------------- !! Agent state not been updated! Agent followed the waypoint left. (rewarded 2.25) 72% of time remaining to reach destination. /------------------- | Step 7 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.25) 68% of time remaining to reach destination. /------------------- | Step 8 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 1.97) 64% of time remaining to reach destination. /------------------- | Step 9 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 2.64) 60% of time remaining to reach destination. /------------------- | Step 10 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -5.33) 56% of time remaining to reach destination. /------------------- | Step 11 Results \------------------- !! Agent state not been updated! Agent drove forward instead of left. (rewarded 0.48) 52% of time remaining to reach destination. /------------------- | Step 12 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08) 48% of time remaining to reach destination. /------------------- | Step 13 Results \------------------- !! Agent state not been updated! Agent drove right instead of left. (rewarded 0.41) 44% of time remaining to reach destination. /------------------- | Step 14 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded 0.89) 40% of time remaining to reach destination. /------------------- | Step 15 Results \------------------- !! Agent state not been updated! Agent properly idled at a red light. (rewarded -0.27) 36% of time remaining to reach destination. /------------------- | Step 16 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.83) 32% of time remaining to reach destination. /------------------- | Step 17 Results \------------------- !! Agent state not been updated! Agent idled at a green light with no oncoming traffic. (rewarded -4.61) 28% of time remaining to reach destination. /------------------- | Step 18 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.67) 24% of time remaining to reach destination. /------------------- | Step 19 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.26) 20% of time remaining to reach destination. /------------------- | Step 20 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41) 16% of time remaining to reach destination. /------------------- | Step 21 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded -0.21) 12% of time remaining to reach destination. /------------------- | Step 22 Results \------------------- !! Agent state not been updated! Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86) 8% of time remaining to reach destination. /------------------- | Step 23 Results \------------------- !! Agent state not been updated! Agent drove forward instead of right. (rewarded 0.12) 4% of time remaining to reach destination. /------------------- | Step 24 Results \------------------- !! Agent state not been updated! Agent attempted driving left through a red light. (rewarded -9.65) 0% of time remaining to reach destination. Trial Aborted! Agent did not reach the destination. Simulation ended. . .
# Load the 'sim_no-learning' log file from the initial simulation results
vs.plot_trials('./sim_no-learning.csv')
Using the visualization above that was produced from your initial simulation, provide an analysis and make several observations about the driving agent. Be sure that you are making at least one observation about each panel present in the visualization. Some things you could consider:
Answer:
The second step to creating an optimized Q-learning driving agent is defining a set of states that the agent can occupy in the environment. Depending on the input, sensory data, and additional variables available to the driving agent, a set of states can be defined for the agent so that it can eventually learn what action it should take when occupying a state. The condition of 'if state then action' for each state is called a policy, and is ultimately what the driving agent is expected to learn. Without defining states, the driving agent would never understand which action is most optimal -- or even what environmental variables and conditions it cares about!
Inspecting the 'build_state()' agent function shows that the driving agent is given the following data from the environment:
'waypoint', which is the direction the Smartcab should drive leading to the destination, relative to the Smartcab's heading.'inputs', which is the sensor data from the Smartcab. It includes 'light', the color of the light.'left', the intended direction of travel for a vehicle to the Smartcab's left. Returns None if no vehicle is present.'right', the intended direction of travel for a vehicle to the Smartcab's right. Returns None if no vehicle is present.'oncoming', the intended direction of travel for a vehicle across the intersection from the Smartcab. Returns None if no vehicle is present.'deadline', which is the number of actions remaining for the Smartcab to reach the destination before running out of time.%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.04)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.21)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.01)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.55)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.18)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.16)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.85)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.22)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.79)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.50)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 2
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.74)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.29)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.41)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.76)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.94)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.58)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.78)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.78)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.10)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.35)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.64)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.56)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
37% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.56)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.94)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.18)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.34)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.47)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.20)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.13)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.88)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 3
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.78)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.40)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.42)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.86)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.11)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.85)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.05)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.56)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.26)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.55)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.54)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.52)
37% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.52)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.03)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
27% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 4
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.35)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.25)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.75)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.07)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 5
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.90)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.97)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.51)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.24)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded -0.32)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.48)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.38)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 6
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.11)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.34)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.69)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.69)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.75)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.13)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.97)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.91)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.17)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.45)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.48)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.41)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 7
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.81)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.61)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.06)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.15)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 8
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded 1.88)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.23)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.20)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.58)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.55)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.07)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.98)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.42)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.92)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.03)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.73)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.00)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.16)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.27)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.33)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.99)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
3% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 9
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.05)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.72)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.12)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.20)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.20)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 10
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.21)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.22)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.62)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.97)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.42)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 11
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove right instead of forward. (rewarded 1.35)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.65)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.45)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.87)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.55)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.72)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.43)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 12
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.85)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.08)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.21)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 0.97)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.75)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -6.00)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.39)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.59)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.96)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 13
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.09)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.59)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.64)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.84)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.44)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.37)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.35)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.96)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.74)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.11)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.13)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.60)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.20)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.45)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.39)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.08)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.22)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.81)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.46)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.14)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.38)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded -0.19)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 14
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.43)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.15)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.02)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.74)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.94)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.40)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 0.57)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 15
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.83)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.10)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.71)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.86)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.34)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.51)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.28)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.34)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.01)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.30)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 16
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.53)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.96)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.79)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.14)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.25)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.73)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.46)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.60)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.44)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.40)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.40)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.54)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.07)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.05)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.76)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 17
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.46)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.55)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.49)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.51)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.34)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 18
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.38)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.88)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.25)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 1.28)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.48)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.86)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.08)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.88)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove right instead of left. (rewarded 1.22)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.56)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.83)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 19
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.07)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -11.00)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.69)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.22)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.35)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.15)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 20
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.04)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.56)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.51)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.43)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.53)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.85)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 1
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.05)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.29)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.17)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.84)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.73)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.84)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.67)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.66)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.83)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 2
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.28)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.73)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.34)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.82)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.48)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 3
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.04)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.78)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded -0.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.52)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.42)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 1.47)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.58)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.47)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 4
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.49)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.16)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.24)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.60)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.89)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 5
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.54)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.56)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.01)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.35)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.27)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 6
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.49)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.70)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.92)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.86)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.32)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.63)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.12)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 7
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.88)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.78)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.93)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded -0.74)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent drove right instead of left. (rewarded -0.54)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 8
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.25)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.55)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.76)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.14)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.61)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.65)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.27)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.82)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.19)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.40)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.69)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.91)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.90)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.35)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.58)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.58)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 9
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.45)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
91% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.96)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.71)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.89)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.47)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded 0.85)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.52)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.58)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.37)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.08)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.26)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.45)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent drove right instead of left. (rewarded 0.41)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.65)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.28)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.45)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.55)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 10
\-------------------------
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
Simulating trial. . .
Agent not set to learn.
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.42)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.28)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.11)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.56)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.23)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.23)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
Simulation ended. . .
Which features available to the agent are most relevant for learning both safety and efficiency? Why are these features appropriate for modeling the Smartcab in the environment? If you did not choose some features, why are those features not appropriate? Please note that whatever features you eventually choose for your agent's state, must be argued for here. That is: your code in agent.py should reflect the features chosen in this answer.
NOTE: You are not allowed to engineer new features for the smartcab.
Answers
Answer:
When defining a set of states that the agent can occupy, it is necessary to consider the size of the state space. That is to say, if you expect the driving agent to learn a policy for each state, you would need to have an optimal action for every state the agent can occupy. If the number of all possible states is very large, it might be the case that the driving agent never learns what to do in some states, which can lead to uninformed decisions. For example, consider a case where the following features are used to define the state of the Smartcab:
('is_raining', 'is_foggy', 'is_red_light', 'turn_left', 'no_traffic', 'previous_turn_left', 'time_of_day').
How frequently would the agent occupy a state like (False, True, True, True, False, False, '3AM')? Without a near-infinite amount of time for training, it's doubtful the agent would ever learn the proper action!
If a state is defined using the features you've selected from Question 4, what would be the size of the state space? Given what you know about the environment and how it is simulated, do you think the driving agent could learn a policy for each possible state within a reasonable number of training trials?
Hint: Consider the combinations of features to calculate the total number of states!
Answer:
For your second implementation, navigate to the 'build_state()' agent function. With the justification you've provided in Question 4, you will now set the 'state' variable to a tuple of all the features necessary for Q-Learning. Confirm your driving agent is updating its state by running the agent file and simulation briefly and note whether the state is displaying. If the visual simulation is used, confirm that the updated state corresponds with what is seen in the simulation.
Note: Remember to reset simulation flags to their default setting when making this observation!
The third step to creating an optimized Q-Learning agent is to begin implementing the functionality of Q-Learning itself. The concept of Q-Learning is fairly straightforward: For every state the agent visits, create an entry in the Q-table for all state-action pairs available. Then, when the agent encounters a state and performs an action, update the Q-value associated with that state-action pair based on the reward received and the iterative update rule implemented. Of course, additional benefits come from Q-Learning, such that we can have the agent choose the best action for each state based on the Q-values of each state-action pair possible. For this project, you will be implementing a decaying, $\epsilon$-greedy Q-learning algorithm with no discount factor. Follow the implementation instructions under each TODO in the agent functions.
Note that the agent attribute self.Q is a dictionary: This is how the Q-table will be formed. Each state will be a key of the self.Q dictionary, and each value will then be another dictionary that holds the action and Q-value. Here is an example:
{ 'state-1': {
'action-1' : Qvalue-1,
'action-2' : Qvalue-2,
...
},
'state-2': {
'action-1' : Qvalue-1,
...
},
...
}
Furthermore, note that you are expected to use a decaying $\epsilon$ (exploration) factor. Hence, as the number of trials increases, $\epsilon$ should decrease towards 0. This is because the agent is expected to learn from its behavior and begin acting on its learned behavior. Additionally, The agent will be tested on what it has learned after $\epsilon$ has passed a certain threshold (the default threshold is 0.05). For the initial Q-Learning implementation, you will be implementing a linear decaying function for $\epsilon$.
To obtain results from the initial Q-Learning implementation, you will need to adjust the following flags and setup:
'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.'log_metrics' - Set this to True to log the simluation results as a .csv file and the Q-table as a .txt file in /logs/.'n_test' - Set this to '10' to perform 10 testing trials.'learning' - Set this to 'True' to tell the driving agent to use your Q-Learning implementation.In addition, use the following decay function for $\epsilon$:
$$ \epsilon_{t+1} = \epsilon_{t} - 0.05, \hspace{10px}\textrm{for trial number } t$$If you have difficulty getting your implementation to work, try setting the 'verbose' flag to True to help debug. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!
Once you have successfully completed the initial Q-Learning simulation, run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded!
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.14)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.66)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.08)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.79)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.28)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.74)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.39)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.03)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.16)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 2
\-------------------------
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.9000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'right')
Agent followed the waypoint right. (rewarded 1.88)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.77)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.85)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.35)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 0.75)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.98)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.58)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.01)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.43)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 0.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.56)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.48)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 3
\-------------------------
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.62)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.29)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.85)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded -0.38)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.10)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.64)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.88)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 4
\-------------------------
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.8000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.92)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 1.37)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.62)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.86)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.03)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.39)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.81)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.18)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.18)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.01)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.99)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.65)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.40)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 5
\-------------------------
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.14)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.34)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.61)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.99)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.77)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.32)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.89)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 6
\-------------------------
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.7000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.38)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.54)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.87)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.67)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.15)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.06)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.06)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 0.07)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.33)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.36)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.65)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.65)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.85)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.48)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 7
\-------------------------
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.01)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.08)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.49)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.23)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.20)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.65)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.24)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.53)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -10.71)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 8
\-------------------------
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.6000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.00)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.59)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.05)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.78)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.69)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.91)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 9
\-------------------------
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.10)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.41)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.29)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.79)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.20)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.35)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.38)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.57)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.76)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 10
\-------------------------
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.5000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 11
\-------------------------
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.34)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.08)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.90)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
73% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 12
\-------------------------
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.4000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.26)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.97)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.85)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.89)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.64)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.83)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.72)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.37)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.40)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 13
\-------------------------
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.40)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.07)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.33)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.87)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.05)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.17)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.68)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.25)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 14
\-------------------------
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.3000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.89)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.80)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.83)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.44)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.33)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 15
\-------------------------
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.74)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.88)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.43)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.22)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.66)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.23)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.68)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.81)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.58)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 16
\-------------------------
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.2000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.29)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.75)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.55)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.59)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 17
\-------------------------
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.05)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.14)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.25)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.45)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.71)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.56)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.70)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.50)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.26)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 0.98)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.61)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.27)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 18
\-------------------------
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
Simulating trial. . .
epsilon = 0.1000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.73)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded -0.01)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.17)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.37)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.26)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.73)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.03)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.90)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.47)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.42)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.11)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.38)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.92)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.65)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.89)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 19
\-------------------------
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
Simulating trial. . .
epsilon = 0.0500; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.69)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.90)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.40)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.93)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.15)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.79)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.08)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 20
\-------------------------
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
Simulating trial. . .
epsilon = -0.0000; alpha = 0.5000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent drove right instead of left. (rewarded 1.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.31)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.80)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.07)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.94)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.11)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.46)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.75)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.75)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.28)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.10)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.32)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.64)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.47)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.77)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.50)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.09)
3% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 1
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.05)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.12)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.18)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.09)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.47)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.42)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove left instead of right. (rewarded 0.10)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.12)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 2
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.31)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.55)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.78)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 3
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 0.26)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.66)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.93)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.55)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.96)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.64)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.16)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.00)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 4
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.83)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.12)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.85)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.96)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.46)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.31)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.69)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.05)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.68)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded -0.27)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.33)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.33)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 5
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.26)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.57)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.67)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.83)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.90)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.36)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.12)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.15)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.31)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.20)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.38)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 6
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.79)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 7
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.15)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.78)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.67)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.65)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.94)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.79)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 8
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.88)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.18)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.71)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.71)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.17)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.17)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.02)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.80)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.03)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 9
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.56)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.93)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded -0.06)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 10
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 1.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.78)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.13)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
Simulation ended. . .
<matplotlib.figure.Figure at 0x935bb70>
# Load the 'sim_default-learning' file from the default Q-Learning simulation
vs.plot_trials('sim_default-learning.csv')
Using the visualization above that was produced from your default Q-Learning simulation, provide an analysis and make observations about the driving agent like in Question 3. Note that the simulation should have also produced the Q-table in a text file which can help you make observations about the agent's learning. Some additional things you could consider:
Answer:
The third step to creating an optimized Q-Learning agent is to perform the optimization! Now that the Q-Learning algorithm is implemented and the driving agent is successfully learning, it's necessary to tune settings and adjust learning paramaters so the driving agent learns both safety and efficiency. Typically this step will require a lot of trial and error, as some settings will invariably make the learning worse. One thing to keep in mind is the act of learning itself and the time that this takes: In theory, we could allow the agent to learn for an incredibly long amount of time; however, another goal of Q-Learning is to transition from experimenting with unlearned behavior to acting on learned behavior. For example, always allowing the agent to perform a random action during training (if $\epsilon = 1$ and never decays) will certainly make it learn, but never let it act. When improving on your Q-Learning implementation, consider the implications it creates and whether it is logistically sensible to make a particular adjustment.
To obtain results from the initial Q-Learning implementation, you will need to adjust the following flags and setup:
'enforce_deadline' - Set this to True to force the driving agent to capture whether it reaches the destination in time.'update_delay' - Set this to a small value (such as 0.01) to reduce the time between steps in each trial.'log_metrics' - Set this to True to log the simluation results as a .csv file and the Q-table as a .txt file in /logs/.'learning' - Set this to 'True' to tell the driving agent to use your Q-Learning implementation.'optimized' - Set this to 'True' to tell the driving agent you are performing an optimized version of the Q-Learning implementation.Additional flags that can be adjusted as part of optimizing the Q-Learning agent:
'n_test' - Set this to some positive number (previously 10) to perform that many testing trials.'alpha' - Set this to a real number between 0 - 1 to adjust the learning rate of the Q-Learning algorithm.'epsilon' - Set this to a real number between 0 - 1 to adjust the starting exploration factor of the Q-Learning algorithm.'tolerance' - set this to some small value larger than 0 (default was 0.05) to set the epsilon threshold for testing.Furthermore, use a decaying function of your choice for $\epsilon$ (the exploration factor). Note that whichever function you use, it must decay to 'tolerance' at a reasonable rate. The Q-Learning agent will not begin testing until this occurs. Some example decaying functions (for $t$, the number of trials):
You may also use a decaying function for $\alpha$ (the learning rate) if you so choose, however this is typically less common. If you do so, be sure that it adheres to the inequality $0 \leq \alpha \leq 1$.
If you have difficulty getting your implementation to work, try setting the 'verbose' flag to True to help debug. Flags that have been set here should be returned to their default setting when debugging. It is important that you understand what each flag does and how it affects the simulation!
Once you have successfully completed the improved Q-Learning simulation, run the code cell below to visualize the results. Note that log files are overwritten when identical simulations are run, so be careful with what log file is being loaded!
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.70)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.85)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.17)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.80)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 2
\-------------------------
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.50)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.26)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.55)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.88)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.82)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.96)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.20)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.51)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.49)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.81)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.38)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 3
\-------------------------
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.69)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.80)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.15)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.48)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.91)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.35)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.16)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.19)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 4
\-------------------------
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.21)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.90)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.17)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.62)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.51)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.27)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.40)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 5
\-------------------------
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.07)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.46)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.77)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.57)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.29)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.50)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 0.63)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove right instead of left. (rewarded -0.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 6
\-------------------------
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 0.93)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.92)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.44)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.31)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.06)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.56)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.93)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.93)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.08)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.74)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.99)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.20)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 7
\-------------------------
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.20)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.01)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.90)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.55)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.28)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.48)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.78)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.55)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.03)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.47)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.67)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 8
\-------------------------
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.73)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.69)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.94)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.16)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 9
\-------------------------
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.69)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.96)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.25)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.12)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.99)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.92)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.55)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 10
\-------------------------
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.00)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.36)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.24)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.51)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.49)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 11
\-------------------------
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.82)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.77)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.62)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.79)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.38)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.77)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.52)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.02)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 12
\-------------------------
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.19)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.66)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.30)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.99)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.26)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.47)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 13
\-------------------------
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.70)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.21)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.69)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.71)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.69)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.29)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.37)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.79)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.30)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.38)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 14
\-------------------------
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.21)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.69)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.02)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.34)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.72)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.53)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.03)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.50)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.26)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.06)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 15
\-------------------------
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.88)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.35)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.27)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.32)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.00)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.43)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.68)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.55)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.80)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 16
\-------------------------
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.41)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.96)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.46)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.38)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.08)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.67)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 17
\-------------------------
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.00)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.74)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.74)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.35)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.10)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.17)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 18
\-------------------------
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.89)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.43)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.10)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.71)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.69)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.85)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.18)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 19
\-------------------------
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.45)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.54)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.48)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.19)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.39)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.08)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.08)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 20
\-------------------------
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent drove forward instead of right. (rewarded 0.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.70)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.16)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.47)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.15)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 21
\-------------------------
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.75)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.36)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.00)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 22
\-------------------------
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.66)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.95)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.05)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.34)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.70)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.68)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded -0.14)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 23
\-------------------------
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.99)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.37)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.69)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.20)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.35)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.85)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.05)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.68)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.56)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 24
\-------------------------
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.31)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.44)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.76)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.65)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent drove right instead of left. (rewarded 0.58)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.55)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 0.71)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.51)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.73)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 25
\-------------------------
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.96)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.33)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.52)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.62)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.07)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.16)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.77)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.42)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.95)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.04)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.87)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.27)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.47)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.77)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.63)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.62)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.93)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.79)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 26
\-------------------------
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.31)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.75)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.49)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.82)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.00)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 27
\-------------------------
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.24)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'right')
Agent drove right instead of left. (rewarded 0.90)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 28
\-------------------------
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.37)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.34)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.03)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.26)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.36)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 29
\-------------------------
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.52)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.10)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 1.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.62)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.05)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.02)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.86)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.18)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.91)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.41)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.22)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.64)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 30
\-------------------------
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.28)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.51)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.42)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.48)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.96)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.23)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.31)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.00)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.63)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 31
\-------------------------
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.03)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.40)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded -0.15)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.91)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded -0.55)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.19)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 32
\-------------------------
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.41)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.82)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.92)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.04)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.22)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.63)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.65)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.65)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.35)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.40)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.27)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.07)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.71)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.55)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 0.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 33
\-------------------------
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.96)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.74)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.52)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.28)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.29)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.44)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.47)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.74)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.05)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.48)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.55)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -10.95)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 34
\-------------------------
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.54)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.91)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.86)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.07)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.99)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.51)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.71)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded -0.16)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.73)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.44)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.56)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.80)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.07)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.39)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.60)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 35
\-------------------------
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.50)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.26)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.01)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded -0.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.60)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.91)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.52)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.14)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 36
\-------------------------
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.02)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.43)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.30)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.13)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.31)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.05)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.37)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 37
\-------------------------
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded -0.04)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.03)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.72)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded -0.61)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.28)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 38
\-------------------------
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.77)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.10)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.02)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.44)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.18)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.46)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.10)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.54)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.59)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.78)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 39
\-------------------------
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.73)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.78)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.29)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.49)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.69)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.43)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.37)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.72)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.90)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove right instead of left. (rewarded 1.70)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.24)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.14)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.59)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.13)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.00)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.29)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.65)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.77)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 40
\-------------------------
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.13)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.77)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 41
\-------------------------
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.60)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.98)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.39)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.46)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.24)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.32)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.51)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.34)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.75)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.87)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 42
\-------------------------
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent drove right instead of left. (rewarded 0.74)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 43
\-------------------------
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.66)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.90)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.08)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.29)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.61)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.24)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded -0.18)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 44
\-------------------------
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.49)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.36)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.49)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.92)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.64)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.77)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.01)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.33)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.41)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.40)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 45
\-------------------------
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.81)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.52)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.68)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.88)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.56)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.02)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.02)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.02)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.85)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.40)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.19)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.71)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.97)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.97)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.25)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.60)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.35)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 0.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 46
\-------------------------
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.83)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.38)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.19)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.47)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 1.10)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.65)
46% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.65)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.70)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.66)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.61)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.05)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.52)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.78)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.97)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 47
\-------------------------
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.22)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.42)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.03)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.81)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.73)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.48)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.41)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.16)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.10)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.10)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.88)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.05)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 48
\-------------------------
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.27)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.98)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.40)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.90)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.73)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.91)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.79)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.65)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.67)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.42)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.93)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.36)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.67)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.05)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.34)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 49
\-------------------------
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.47)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.20)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.85)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.29)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.75)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.17)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.48)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.28)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.70)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 50
\-------------------------
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.89)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.71)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.85)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.03)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded 1.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.40)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.48)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.49)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.96)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.45)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 51
\-------------------------
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.46)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.49)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.40)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.38)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.97)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.75)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.89)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 52
\-------------------------
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.91)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.43)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.69)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.89)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.83)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.13)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.90)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.08)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 53
\-------------------------
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.91)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.04)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.06)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.63)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.42)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.04)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.04)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.36)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.50)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.95)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.01)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.58)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.70)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.29)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.68)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 54
\-------------------------
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.09)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.17)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.93)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.91)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.03)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.93)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.58)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.10)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.30)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.81)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.18)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded -0.10)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.65)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.49)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.49)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.01)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 55
\-------------------------
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.08)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.73)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.16)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.17)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.54)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.80)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.42)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.73)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.52)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.52)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.20)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 56
\-------------------------
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.19)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.37)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.51)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.38)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.13)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.73)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.92)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 57
\-------------------------
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 1.11)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.54)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.52)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.25)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.01)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.51)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 58
\-------------------------
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.92)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.10)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.05)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.03)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.17)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.44)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.99)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.23)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.44)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.16)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.35)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.35)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.51)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.91)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove right instead of forward. (rewarded -0.56)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 59
\-------------------------
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.59)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.38)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 0.75)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.80)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.96)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.12)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.27)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.05)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.05)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.33)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 60
\-------------------------
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.99)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.25)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.28)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 61
\-------------------------
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.05)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.53)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.30)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.06)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.00)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.49)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.12)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.35)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.60)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.61)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.04)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.11)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.28)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.37)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.46)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.53)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.18)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.64)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.86)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.72)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 62
\-------------------------
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.58)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.32)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 2.78)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.35)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.06)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.11)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.65)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.04)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.52)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 63
\-------------------------
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.24)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.51)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.05)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.59)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.32)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.65)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.07)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.38)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.90)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 64
\-------------------------
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.20)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.71)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.10)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.95)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.13)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.02)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.31)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.00)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.86)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.14)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.42)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.53)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.82)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 65
\-------------------------
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.38)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.20)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.52)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.12)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 0.95)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.97)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.49)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.59)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.07)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.26)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.70)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.37)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.53)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.17)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 66
\-------------------------
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.30)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.63)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.95)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.46)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.60)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.74)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.29)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.55)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.53)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.92)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.19)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.30)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.97)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 67
\-------------------------
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.74)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.20)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.20)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.75)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.82)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.57)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.59)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.29)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.92)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.17)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.20)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 68
\-------------------------
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.94)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.10)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.18)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.41)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 69
\-------------------------
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.79)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.10)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.47)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.03)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.49)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.98)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.08)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.74)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.69)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.93)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.44)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.42)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.50)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.36)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.15)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.50)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.27)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.15)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.48)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 70
\-------------------------
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.52)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.93)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.69)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.43)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.81)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.59)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.64)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove right instead of forward. (rewarded -0.59)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.40)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 71
\-------------------------
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.31)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.58)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.87)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.86)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.93)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.21)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.69)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.38)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.94)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded -0.44)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.40)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.40)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove forward instead of right. (rewarded 0.91)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 72
\-------------------------
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.95)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.91)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.77)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.13)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.33)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.12)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.12)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.92)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.21)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.38)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.26)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.04)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.40)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.41)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 73
\-------------------------
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.06)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.87)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.78)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.95)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.98)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.24)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.68)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 74
\-------------------------
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.97)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.75)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.32)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.07)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.71)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.23)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.67)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.25)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.36)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 75
\-------------------------
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.84)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.52)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.37)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.48)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.18)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.04)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.82)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 76
\-------------------------
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.59)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 1.20)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.63)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.13)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.29)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.71)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.48)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.53)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.31)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.61)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.45)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.76)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.07)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded -0.05)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.23)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 77
\-------------------------
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded -0.02)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.61)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.76)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.33)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.13)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.07)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.82)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.55)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 78
\-------------------------
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.45)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.40)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.30)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.69)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.17)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.70)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.85)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.03)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.69)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.50)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.51)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.53)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.78)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded -0.46)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded -0.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 79
\-------------------------
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.36)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.77)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.80)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.76)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.65)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.61)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.27)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.57)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.00)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.06)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.81)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.08)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.36)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 80
\-------------------------
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.01)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.21)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.29)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.42)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.98)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.76)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.55)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.10)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.95)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 81
\-------------------------
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.64)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.06)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.65)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.08)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.17)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 82
\-------------------------
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.43)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.02)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.84)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.08)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.66)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.80)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.91)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.94)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.28)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.17)
23% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 83
\-------------------------
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.85)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.02)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.64)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.21)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.67)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.24)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.56)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.27)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.67)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.27)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.44)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.44)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.12)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 84
\-------------------------
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1901; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.42)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.33)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.09)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.45)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.04)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.62)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded -0.06)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.59)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.88)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.14)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 85
\-------------------------
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1864; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'right')
Agent drove right instead of left. (rewarded 0.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.40)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.47)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.68)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.86)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.00)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 86
\-------------------------
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1827; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.69)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.83)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.63)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.96)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.16)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.26)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.78)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.80)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.67)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.32)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.40)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.38)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.24)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.37)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.46)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.99)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.57)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.34)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 87
\-------------------------
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1791; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.78)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.57)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.54)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.75)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.44)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.57)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.26)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 88
\-------------------------
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1755; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.94)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.23)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.39)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.91)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.83)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.50)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 89
\-------------------------
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1720; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.04)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.91)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.06)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.80)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 90
\-------------------------
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1686; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.94)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.29)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.45)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.08)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.47)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.35)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.83)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 91
\-------------------------
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1653; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.07)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.84)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.28)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 92
\-------------------------
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1620; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.45)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.09)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.09)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.87)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.30)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.50)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.59)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.53)
63% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 93
\-------------------------
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1588; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.74)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.81)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.26)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.37)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.85)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.37)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.34)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.80)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.24)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.63)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.65)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.09)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.15)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.17)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.96)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.77)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 94
\-------------------------
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1557; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.98)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.58)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.39)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.12)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.42)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.49)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.49)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.45)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 95
\-------------------------
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1526; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.52)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded -0.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.29)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.78)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.13)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.33)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.00)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.17)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 96
\-------------------------
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1496; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.21)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.96)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.35)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.26)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.34)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.05)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded -0.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.11)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 97
\-------------------------
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1466; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.54)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.55)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 98
\-------------------------
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1437; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent drove forward instead of right. (rewarded 1.09)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.19)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.70)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.20)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.46)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.40)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.50)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.23)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.48)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.79)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 99
\-------------------------
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1409; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.29)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.91)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.60)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.93)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.08)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.64)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.68)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.48)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.95)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.04)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.46)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 100
\-------------------------
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1381; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.82)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.48)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.40)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.69)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.19)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.48)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.00)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.70)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.37)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.70)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.22)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.09)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 101
\-------------------------
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1353; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.21)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.78)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.85)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.51)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.55)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.73)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 1.17)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.41)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.62)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent drove right instead of forward. (rewarded 1.12)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.74)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.66)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.44)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.73)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.53)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.95)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.33)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.43)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.94)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 102
\-------------------------
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1327; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.54)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 0.71)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.35)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 103
\-------------------------
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1300; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.48)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.30)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.86)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.59)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.60)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.10)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'right')
Agent drove right instead of left. (rewarded 0.07)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.52)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.12)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.91)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.88)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 104
\-------------------------
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1275; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.44)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.39)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.33)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 105
\-------------------------
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1249; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.34)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.70)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.67)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.99)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.35)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.56)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.70)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.47)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.00)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.04)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.32)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.69)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.69)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 0.18)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 106
\-------------------------
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1225; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.50)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.62)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.13)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.40)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.14)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.67)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.82)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.78)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.97)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.02)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.26)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.48)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.15)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 107
\-------------------------
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1200; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.61)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.61)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.21)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.22)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.37)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.48)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.12)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.88)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 108
\-------------------------
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1177; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.38)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.52)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 0.45)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.78)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.48)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 109
\-------------------------
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1153; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.33)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.43)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.25)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.48)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.41)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.60)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 110
\-------------------------
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1130; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.78)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.07)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.17)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.40)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.03)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.62)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent followed the waypoint forward. (rewarded 1.13)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.31)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.97)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.83)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.31)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.58)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.75)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 111
\-------------------------
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1108; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.85)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.03)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.52)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.80)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.80)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.41)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 112
\-------------------------
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1086; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.34)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.57)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.07)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.59)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded -0.07)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.85)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 113
\-------------------------
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1065; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.91)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.19)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.16)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.26)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.28)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.99)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.76)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 114
\-------------------------
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1044; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.75)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.59)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.35)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.85)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.24)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.92)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.23)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.03)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.41)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.65)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded -0.73)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 115
\-------------------------
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1023; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.59)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.40)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.09)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.20)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.34)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.32)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.06)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 116
\-------------------------
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
Simulating trial. . .
epsilon = 0.1003; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.64)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.69)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.33)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.23)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.65)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.32)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.70)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.25)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.28)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.51)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.23)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 117
\-------------------------
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0983; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.00)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.71)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 118
\-------------------------
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0963; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.67)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.53)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.79)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.56)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.26)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.08)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.87)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.32)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.66)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.84)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.18)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 119
\-------------------------
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0944; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.64)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.84)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.48)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.45)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.31)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.34)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.01)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 120
\-------------------------
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0926; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.50)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.95)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.45)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.02)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 121
\-------------------------
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0907; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.42)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.59)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.72)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.42)
53% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 122
\-------------------------
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0889; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.71)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.82)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.44)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.58)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded -0.08)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.03)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.23)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.56)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.32)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.65)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.13)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.15)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.05)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 123
\-------------------------
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0872; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.94)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.86)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.34)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 1.60)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.14)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.43)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.78)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 124
\-------------------------
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0854; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.77)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.52)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.74)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.02)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.82)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 125
\-------------------------
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0837; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.95)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.11)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.17)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 126
\-------------------------
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0821; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.77)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.38)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.61)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.90)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.29)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.05)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.04)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.79)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.13)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.74)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
37% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 127
\-------------------------
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0805; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.99)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.83)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.45)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.32)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 128
\-------------------------
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0789; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.24)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.38)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.67)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.86)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.21)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.98)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.76)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.73)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.25)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.57)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.45)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.40)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.05)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.05)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 129
\-------------------------
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0773; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.00)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.75)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.74)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.99)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.83)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.95)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded -0.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.01)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.17)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 130
\-------------------------
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0758; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 0.91)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.74)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.30)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.65)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.89)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.77)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.59)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.15)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.49)
23% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 131
\-------------------------
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0743; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.10)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.72)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.95)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.61)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.03)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.47)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.08)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.68)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded 1.39)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.65)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.09)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.35)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.47)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.64)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.97)
34% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.97)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.42)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.81)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.62)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.71)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.60)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.52)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.96)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 132
\-------------------------
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0728; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.80)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.81)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.02)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.24)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.16)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.10)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.83)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.31)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.20)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.19)
13% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 133
\-------------------------
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0714; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.70)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.19)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.56)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.77)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.03)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 134
\-------------------------
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0699; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.16)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.87)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.87)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.52)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.51)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.56)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 135
\-------------------------
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0686; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.22)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.44)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.20)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.08)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.55)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.18)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.74)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 136
\-------------------------
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0672; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.17)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.82)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.68)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.67)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.10)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.37)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded -0.65)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 137
\-------------------------
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0659; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.56)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.57)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.50)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.54)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.58)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded -0.19)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.96)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 138
\-------------------------
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0646; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.29)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.74)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.81)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.80)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.69)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.03)
63% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 139
\-------------------------
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0633; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent followed the waypoint right. (rewarded 2.73)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.72)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.51)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.64)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.93)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.20)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.05)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.20)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.60)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.26)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.13)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.28)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.87)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.31)
14% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.31)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.26)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.11)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.31)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 140
\-------------------------
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0620; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.74)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.28)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.00)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 141
\-------------------------
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0608; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.20)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.40)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.56)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 142
\-------------------------
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0596; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.22)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.14)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.54)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 143
\-------------------------
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0584; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.59)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.87)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded 0.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.05)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.80)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.82)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.20)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.32)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.67)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 144
\-------------------------
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0573; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.45)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.50)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.60)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.62)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.80)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.39)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.25)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.27)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.03)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.07)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded -0.28)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.00)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.07)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.99)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.46)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.28)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 145
\-------------------------
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0561; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.07)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.92)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.32)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.09)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.17)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.18)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.69)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.66)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.70)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.63)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.09)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.38)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.19)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.40)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.03)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.82)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.76)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.37)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.35)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.63)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 146
\-------------------------
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0550; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.93)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.39)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 0.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.48)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.63)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.01)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.32)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.61)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.57)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.42)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 147
\-------------------------
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0539; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.38)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.62)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.12)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.07)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.70)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.92)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.63)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.44)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.01)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.65)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.65)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded -0.63)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.71)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 148
\-------------------------
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0529; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 1.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.60)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.25)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.54)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 149
\-------------------------
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0518; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.50)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.74)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.77)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.31)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded -0.19)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.72)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.56)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.46)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 150
\-------------------------
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0508; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.44)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 151
\-------------------------
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
Simulating trial. . .
epsilon = 0.0498; alpha = 0.0200
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.78)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.60)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 2.21)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 1
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.16)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.66)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.23)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.67)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.16)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.82)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.26)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.72)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.30)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.69)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent drove right instead of forward. (rewarded -0.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 2
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.26)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.66)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.58)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 3
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.17)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.64)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.58)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.58)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.87)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.39)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 4
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.03)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.63)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.15)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.22)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.31)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.00)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.52)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 5
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.26)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.38)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.37)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Testing trial 6
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.93)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 1.57)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.97)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded -0.00)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.11)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.97)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.06)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.28)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.01)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.46)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 7
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.56)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.07)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.84)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.26)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.63)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.88)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.19)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 8
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.92)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.32)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.28)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.51)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.46)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.65)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.22)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.86)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.02)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.02)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.09)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.61)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.20)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.47)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.62)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.13)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.47)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.47)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.57)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.81)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.65)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 9
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.09)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.37)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.02)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.37)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.42)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.65)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.82)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Testing trial 10
\-------------------------
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
Simulating trial. . .
epsilon = 0.0000; alpha = 0.0000
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.28)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.24)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.14)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.60)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.49)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.10)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.47)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.35)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
Simulation ended. . .
<matplotlib.figure.Figure at 0x5bbb080>
# Load the 'sim_improved-learning' file from the improved Q-Learning simulation
vs.plot_trials('sim_improved-learning.csv')
print "decay function is :ϵ=e−at, alpha=0.02,epsilon=0.5"
decay function is :ϵ=e−at, alpha=0.02,epsilon=0.5
Using the visualization above that was produced from your improved Q-Learning simulation, provide a final analysis and make observations about the improved driving agent like in Question 6. Questions you should answer:
%run smartcab/agent.py
/-------------------------
| Training trial 1
\-------------------------
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
Simulating trial. . .
epsilon = 1.0000; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.38)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.96)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.98)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.62)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.06)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.74)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.26)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.38)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
34% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
31% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.65)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.62)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
6% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.39)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 2
\-------------------------
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9980; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.91)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.77)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.85)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.95)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.46)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove left instead of right. (rewarded 0.92)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.41)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.37)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.14)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.10)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.82)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.50)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 3
\-------------------------
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9960; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.76)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.57)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.72)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.72)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.24)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.97)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.63)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 4
\-------------------------
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9940; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.62)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.61)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.13)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.83)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.68)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.95)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.80)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.33)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded -0.03)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 5
\-------------------------
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9920; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.99)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.56)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.18)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.28)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.22)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.22)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.93)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.05)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.47)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 6
\-------------------------
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9900; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.85)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.71)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.55)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.67)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.40)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.71)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.62)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.86)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded -0.18)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.48)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.74)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.28)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.18)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.35)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.17)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded -0.07)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.90)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.10)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 7
\-------------------------
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9881; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.74)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.65)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.33)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.84)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.15)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.99)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 8
\-------------------------
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9861; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.71)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.97)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.94)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.29)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 9
\-------------------------
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9841; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.04)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.57)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.57)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.19)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.75)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.56)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.85)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.07)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.76)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.36)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.49)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.62)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.99)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 10
\-------------------------
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9822; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.84)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.19)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.18)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.48)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.12)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.36)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.57)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.53)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 11
\-------------------------
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9802; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.00)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.25)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.80)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.26)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.00)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.26)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.89)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.88)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.56)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.08)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 12
\-------------------------
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9782; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.34)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.85)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.82)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.45)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 13
\-------------------------
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9763; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.84)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.97)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 14
\-------------------------
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9743; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.47)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.14)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.47)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.18)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.65)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.07)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.90)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.51)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.03)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.04)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.47)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.56)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.18)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.29)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 15
\-------------------------
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9724; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.42)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.15)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.70)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.78)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 0.98)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.40)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.54)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.92)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.35)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.90)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.13)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.35)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.35)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.68)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.13)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.13)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.70)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.03)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.28)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.80)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.57)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.81)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 16
\-------------------------
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9704; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.86)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.96)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.13)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.78)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.80)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 17
\-------------------------
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9685; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.40)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.35)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.23)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.97)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.55)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.09)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.19)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.28)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.38)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.57)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.25)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 18
\-------------------------
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9666; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.78)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.53)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.87)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded -0.09)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.45)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.49)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.94)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.43)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.08)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 19
\-------------------------
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9646; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.49)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.98)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 0.32)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.26)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 0.79)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 0.43)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.88)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 20
\-------------------------
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9627; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.15)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.78)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.94)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.05)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.19)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.02)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.01)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.37)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 21
\-------------------------
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9608; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 1.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.16)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.40)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.01)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 22
\-------------------------
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9589; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.42)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.30)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.03)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.19)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.13)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.61)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.73)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.21)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.73)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.55)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 23
\-------------------------
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9570; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.20)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.29)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.79)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.82)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.75)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.62)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.71)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.20)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.28)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.22)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.51)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.16)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.26)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.50)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.39)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 24
\-------------------------
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9550; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.74)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.56)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 0.06)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded -0.31)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.85)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.86)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 25
\-------------------------
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9531; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.83)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.46)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.57)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.01)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.37)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.37)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.85)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 26
\-------------------------
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9512; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.94)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.24)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.77)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 27
\-------------------------
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9493; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.93)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.93)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.14)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.16)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.94)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.92)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.56)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.92)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 28
\-------------------------
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9474; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.93)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.64)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.99)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.49)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 1.12)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.93)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.69)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.13)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.13)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded -0.13)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.04)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.92)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.18)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.14)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.12)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.02)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 0.11)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.54)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 29
\-------------------------
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9455; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.82)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.41)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.76)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove forward instead of right. (rewarded 0.37)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.23)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.92)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.94)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.40)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.64)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 30
\-------------------------
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9436; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.22)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.47)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 1.65)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove left instead of right. (rewarded 0.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.59)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.67)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.74)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 31
\-------------------------
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9418; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.74)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.52)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.54)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.74)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.51)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.49)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.13)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.27)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.19)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.58)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.42)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent drove right instead of forward. (rewarded -0.30)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 32
\-------------------------
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9399; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.90)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.13)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.26)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.40)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.09)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 0.41)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.97)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.71)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 33
\-------------------------
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9380; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 0.73)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.93)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.76)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.43)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.76)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.14)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.37)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.24)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded -0.43)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.46)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.72)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 34
\-------------------------
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9361; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 1.97)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.87)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.04)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.86)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.14)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.13)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.32)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.22)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.34)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.25)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.59)
63% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 35
\-------------------------
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9343; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.92)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.31)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.59)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.87)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.80)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.44)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.61)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.32)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.65)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 36
\-------------------------
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9324; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.48)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.35)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.44)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded -0.59)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.50)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 37
\-------------------------
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9305; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.62)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.35)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.48)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.79)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 0.83)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.80)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.03)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.46)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.90)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.46)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.04)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 38
\-------------------------
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9287; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.32)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.05)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.14)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.62)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.96)
63% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 39
\-------------------------
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9268; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.92)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.03)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.65)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 40
\-------------------------
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9250; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.02)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.96)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.96)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.92)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.20)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.19)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.18)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.27)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.08)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 41
\-------------------------
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9231; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.42)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.98)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.30)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.35)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.00)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 0.38)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.73)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 42
\-------------------------
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9213; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.95)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.80)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.25)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.74)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.46)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.09)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.00)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 43
\-------------------------
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9194; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.88)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.49)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.13)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.93)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent drove forward instead of right. (rewarded 0.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.23)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 44
\-------------------------
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9176; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.11)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.84)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.24)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.82)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.79)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.21)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.81)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.81)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 45
\-------------------------
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9158; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.57)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.16)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.19)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.71)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.17)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.57)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 46
\-------------------------
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9139; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.54)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.19)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.02)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.14)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.07)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.53)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.45)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.35)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.91)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.49)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.44)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.97)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.48)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.53)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 47
\-------------------------
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9121; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.80)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.90)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.02)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.49)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.59)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.23)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.29)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.41)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.69)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.15)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.64)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.04)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.70)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.75)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.94)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.59)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.24)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.53)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded -0.14)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 48
\-------------------------
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9103; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.40)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.32)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.19)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.05)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.79)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.15)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.76)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.01)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.89)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.91)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.11)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.64)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 49
\-------------------------
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9085; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.37)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.41)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.37)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.20)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.00)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.93)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.69)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 50
\-------------------------
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9066; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.79)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.85)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.15)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.10)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 0.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 51
\-------------------------
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9048; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.44)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.94)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.68)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 52
\-------------------------
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9030; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.72)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.75)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.74)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.70)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.60)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.41)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.77)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.77)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 53
\-------------------------
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.9012; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.40)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.69)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.08)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.63)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.76)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent followed the waypoint left. (rewarded 0.68)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.69)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.26)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.95)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.48)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 54
\-------------------------
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8994; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.30)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.80)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.48)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.92)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.85)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.04)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.30)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.10)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.53)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.19)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 0.12)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.53)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.21)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.70)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.30)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.41)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 55
\-------------------------
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8976; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded -0.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.32)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.36)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.97)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.66)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.29)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.84)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.39)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 56
\-------------------------
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8958; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.43)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.47)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.01)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.91)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.14)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 57
\-------------------------
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8940; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.11)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.56)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.85)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 0.18)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.74)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.64)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.96)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.83)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.62)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded -0.12)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 58
\-------------------------
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8923; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.07)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.43)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 59
\-------------------------
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8905; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.39)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.21)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.93)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded -0.09)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.51)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.10)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.66)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.87)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.87)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.96)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.09)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.95)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.26)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.39)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 60
\-------------------------
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8887; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.63)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded 0.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.65)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 0.92)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 61
\-------------------------
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8869; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.64)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.07)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.89)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.14)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.54)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.39)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.92)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.66)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.00)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.52)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.24)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.10)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.29)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 62
\-------------------------
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8851; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.24)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.62)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.55)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.34)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.81)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.81)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.17)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.12)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.64)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.57)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.40)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.88)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.46)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.76)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.84)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 63
\-------------------------
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8834; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.52)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.74)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.35)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.29)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.54)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded -0.14)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.34)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.07)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.15)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 64
\-------------------------
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8816; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.53)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.65)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 65
\-------------------------
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8799; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.62)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.51)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.22)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.42)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.03)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded -0.61)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.16)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.03)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 66
\-------------------------
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8781; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.81)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.29)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.26)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.32)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.16)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.57)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.63)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 67
\-------------------------
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8763; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.83)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.94)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 68
\-------------------------
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8746; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.74)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.16)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.23)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.17)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.01)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.07)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.34)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 69
\-------------------------
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8728; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.42)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.53)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.14)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.03)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.48)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.75)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.21)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.18)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.18)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.63)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.53)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.79)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.58)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.13)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 70
\-------------------------
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8711; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.25)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.13)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.74)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.74)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.88)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.13)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.72)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.45)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 71
\-------------------------
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8694; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.58)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.91)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.58)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.11)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded -0.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 1.19)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.46)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 72
\-------------------------
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8676; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.11)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.36)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.34)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.85)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.72)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.35)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.81)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.51)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.66)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.15)
46% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.15)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.06)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.57)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.17)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
29% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 73
\-------------------------
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8659; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.87)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.29)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded -0.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.57)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.83)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 74
\-------------------------
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8642; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.84)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.13)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.58)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.32)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.43)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.51)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 75
\-------------------------
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8624; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.00)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.52)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.46)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.32)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.32)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.31)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.90)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 76
\-------------------------
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8607; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.20)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.68)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.92)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.19)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.39)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.42)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.75)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.51)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.05)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.31)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.50)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.72)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.77)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded -0.22)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.65)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 77
\-------------------------
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8590; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.74)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.83)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.02)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.70)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.35)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.35)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.90)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 78
\-------------------------
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8573; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove left instead of right. (rewarded 0.90)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.53)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.08)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.68)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.28)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.90)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.29)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.73)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.91)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.53)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.25)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.27)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.10)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.36)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 79
\-------------------------
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8556; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.45)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.12)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.45)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.96)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.21)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.03)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.48)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.74)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 80
\-------------------------
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8538; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.45)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.31)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.70)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.14)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.73)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.37)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.99)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.73)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.51)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.00)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.78)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.18)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.35)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 81
\-------------------------
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8521; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.33)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.72)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.96)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.00)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.14)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.21)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.72)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.77)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.16)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.48)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.02)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.84)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.87)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 82
\-------------------------
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8504; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.66)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.39)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.44)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.77)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.77)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.39)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.46)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.08)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.71)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 83
\-------------------------
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8487; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.67)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.75)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.75)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.78)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.59)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.18)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.97)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.73)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.62)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.23)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 84
\-------------------------
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8470; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.29)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -41.00)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded -0.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.54)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.30)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 85
\-------------------------
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8454; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.30)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.49)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.06)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.26)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.83)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.96)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.92)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.60)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.82)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 86
\-------------------------
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8437; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.71)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.99)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.95)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.23)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.71)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded -0.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.41)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded -0.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.65)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 87
\-------------------------
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8420; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.80)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.15)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.63)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.57)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.38)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.15)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.89)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 88
\-------------------------
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8403; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.58)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.58)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.84)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.39)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.54)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded -0.39)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.96)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 89
\-------------------------
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8386; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.10)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.57)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.34)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.73)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.26)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.49)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 90
\-------------------------
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8369; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.49)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.74)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.33)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.21)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.61)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.63)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.24)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.67)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.26)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.69)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.47)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.12)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 91
\-------------------------
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8353; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.61)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.72)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.83)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.68)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.12)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.17)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.10)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.40)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.11)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 92
\-------------------------
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8336; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.91)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.82)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.73)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.77)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.18)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.65)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded -0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 93
\-------------------------
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8319; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.56)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.57)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.17)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.52)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.18)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.57)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.71)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 94
\-------------------------
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8303; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.52)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.33)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.58)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.56)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.13)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.79)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove left instead of right. (rewarded 0.74)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.44)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.24)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.17)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent properly idled at a red light. (rewarded 0.44)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.05)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 95
\-------------------------
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8286; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.40)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.11)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.18)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 0.85)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.76)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.01)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.10)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.50)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 96
\-------------------------
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8270; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.98)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.37)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.04)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.22)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.84)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.63)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.51)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.11)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.91)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.78)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.45)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.87)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 97
\-------------------------
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8253; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.28)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.35)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.11)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.68)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.55)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent drove forward instead of right. (rewarded -0.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.88)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.14)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 98
\-------------------------
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8237; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.40)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.83)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.22)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.30)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.08)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.34)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.94)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.32)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.83)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.36)
27% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 99
\-------------------------
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8220; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 100
\-------------------------
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8204; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.84)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.54)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.77)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.23)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.80)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.08)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.28)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.28)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.80)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 101
\-------------------------
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8187; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.61)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.90)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 0.03)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.52)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.47)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.19)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.85)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded 0.85)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.66)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.95)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 102
\-------------------------
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8171; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.58)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.64)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.50)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.63)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.27)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.79)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.14)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.64)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 103
\-------------------------
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8155; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.80)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.23)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.15)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.42)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 104
\-------------------------
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8138; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.97)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.94)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.46)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.05)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.14)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.26)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.58)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.42)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.54)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.07)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.60)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.77)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 105
\-------------------------
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8122; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.66)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.22)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 0.15)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.16)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.06)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.22)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.62)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.06)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 106
\-------------------------
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8106; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.82)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.39)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.77)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.06)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.25)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.81)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.89)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.89)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.16)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.02)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.87)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.76)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.37)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.92)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 107
\-------------------------
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8090; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 0.09)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.40)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.03)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.23)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.01)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.28)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.21)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.26)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.97)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.31)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.29)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.05)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.99)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.04)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.38)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 108
\-------------------------
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8073; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.64)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.95)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 109
\-------------------------
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8057; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.96)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.19)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.17)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.87)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 110
\-------------------------
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8041; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.86)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.02)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.02)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 0.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.06)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.89)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.56)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.73)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 0.66)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.96)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.46)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 111
\-------------------------
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8025; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.05)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.60)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.45)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.56)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.22)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.88)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.96)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 1.32)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.14)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.85)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.98)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 112
\-------------------------
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.8009; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.16)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.49)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.23)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.83)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 113
\-------------------------
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7993; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.01)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.66)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.76)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.76)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.11)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.77)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.54)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.93)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.39)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.69)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.60)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
27% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 114
\-------------------------
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7977; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.15)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.74)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.59)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.05)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.84)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.51)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.42)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.31)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.26)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.15)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.88)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.25)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.29)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.28)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.57)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.89)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 115
\-------------------------
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7961; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.83)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.80)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 0.99)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.31)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.16)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.87)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.06)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.24)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.63)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.07)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.36)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.45)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.34)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.75)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.26)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.58)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.54)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.93)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 116
\-------------------------
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7945; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.32)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.28)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.34)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.87)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.79)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.56)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 117
\-------------------------
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7929; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.74)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -6.00)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 1.65)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.03)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.06)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.08)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.63)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.46)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.16)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.77)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.42)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.12)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.91)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.53)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.91)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.53)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.97)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.21)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.77)
29% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 118
\-------------------------
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7914; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.26)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.67)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.77)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.42)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.43)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.39)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.57)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.71)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.24)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.43)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.68)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.60)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.24)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.37)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.14)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.28)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.45)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.33)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.94)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.72)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded -0.58)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 119
\-------------------------
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7898; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.19)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.07)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.33)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.17)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.15)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.84)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.14)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.43)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded -0.00)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.54)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.03)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.61)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.35)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.67)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.98)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.61)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 120
\-------------------------
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7882; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 2.23)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.97)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.59)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.51)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.10)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.54)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.42)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.92)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.37)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.38)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.38)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.99)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 121
\-------------------------
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7866; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.84)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.77)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.22)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded -0.08)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.43)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded -0.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.48)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 122
\-------------------------
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7851; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.18)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.56)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.69)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.69)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.19)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.48)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.19)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.09)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove left instead of forward. (rewarded -0.21)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.84)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded -0.16)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.48)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.95)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.35)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.26)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.26)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.18)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.68)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.77)
7% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.77)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.09)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.09)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 123
\-------------------------
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7835; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.01)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.89)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.71)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 124
\-------------------------
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7819; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.80)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.94)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.63)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.27)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.42)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.18)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 125
\-------------------------
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7804; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.06)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.63)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 0.80)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded -0.06)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.02)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.22)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.01)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.40)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.40)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.31)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.31)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.21)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 126
\-------------------------
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7788; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.87)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.71)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.44)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.98)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.22)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.99)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.84)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.46)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.14)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent drove right instead of forward. (rewarded 0.69)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 127
\-------------------------
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7772; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.92)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.28)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.83)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.97)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.98)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.94)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.41)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.56)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.25)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.79)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 128
\-------------------------
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7757; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.45)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.07)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.61)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.57)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.15)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.45)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.71)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.18)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 129
\-------------------------
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7741; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.18)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.98)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.10)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove left instead of forward. (rewarded 1.39)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded -0.44)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.00)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 130
\-------------------------
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7726; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.37)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.14)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.45)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.90)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.92)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.26)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.28)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.88)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.56)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.72)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.79)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 131
\-------------------------
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7711; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.84)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 0.12)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.26)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.68)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.41)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.96)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.19)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.39)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.84)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.50)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.20)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.77)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 132
\-------------------------
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7695; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.57)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.41)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.31)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.33)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.52)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 133
\-------------------------
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7680; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.58)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.65)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.61)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.93)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.24)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 0.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded 0.17)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.84)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.84)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.75)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.87)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.71)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.63)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.19)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.81)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 134
\-------------------------
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7664; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.16)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.86)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.78)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.44)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.50)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.36)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.21)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.26)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.88)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 135
\-------------------------
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7649; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.92)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.43)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 0.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.78)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.02)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.82)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.98)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.32)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.60)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 136
\-------------------------
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7634; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.06)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.73)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.31)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.01)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.56)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.92)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.62)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.06)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.45)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.76)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.37)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.52)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.64)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.09)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded -0.14)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.35)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.10)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded -0.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 137
\-------------------------
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7619; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.93)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.55)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.74)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.74)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.75)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.93)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.79)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.79)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.61)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.52)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.32)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.67)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.30)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.57)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.51)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.82)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded -0.29)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded -0.78)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 138
\-------------------------
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7603; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.99)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.03)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded -0.25)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.25)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.82)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.31)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.32)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.66)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.48)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 139
\-------------------------
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7588; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.40)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.67)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 140
\-------------------------
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7573; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.05)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.75)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.36)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.98)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.15)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.24)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.10)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 141
\-------------------------
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7558; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.66)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.22)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.58)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.62)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.62)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.87)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded -0.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 142
\-------------------------
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7543; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.14)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.47)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.42)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.22)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.29)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.42)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.00)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.01)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.73)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.81)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.05)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 143
\-------------------------
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7528; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.44)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.52)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.36)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 0.71)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.54)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.87)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.87)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.01)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.82)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.35)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.98)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.18)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.87)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 144
\-------------------------
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7513; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.32)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.05)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.05)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.61)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.42)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.55)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.18)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.34)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.38)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.54)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.58)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.48)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded -0.63)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 145
\-------------------------
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7498; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.95)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.17)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.58)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.41)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 1.10)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.35)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded -0.13)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.32)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.12)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.35)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.60)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.52)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 146
\-------------------------
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7483; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.55)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.67)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.92)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent properly idled at a red light. (rewarded 1.33)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.43)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.51)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.48)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 147
\-------------------------
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7468; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.87)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.21)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.60)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.36)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent drove right instead of left. (rewarded 1.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.09)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.57)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 0.71)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.68)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.03)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 148
\-------------------------
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7453; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.42)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.82)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.49)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.73)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.03)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.80)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 149
\-------------------------
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7438; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.38)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 150
\-------------------------
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7423; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.67)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.76)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.71)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.18)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.53)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.02)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.66)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.01)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.07)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.50)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.57)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.23)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.31)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 151
\-------------------------
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7408; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.43)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.79)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.73)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.55)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.21)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.56)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.50)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.44)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.65)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.59)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 1.32)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.81)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.74)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.53)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 152
\-------------------------
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7393; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.41)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.58)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.31)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.49)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.49)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.43)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.07)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.73)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.12)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.09)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 153
\-------------------------
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7379; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.71)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.00)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.65)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.99)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.97)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 154
\-------------------------
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7364; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.30)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.32)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 155
\-------------------------
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7349; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.87)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.68)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.81)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.92)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.58)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.42)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 156
\-------------------------
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7334; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.77)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.58)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.26)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.89)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.43)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.56)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.01)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.78)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.73)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.57)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 157
\-------------------------
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7320; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 0.29)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.68)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.85)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.84)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.95)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.94)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.15)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.27)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.27)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.62)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.62)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.58)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.32)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.06)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.36)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.06)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent drove forward instead of right. (rewarded -0.33)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.42)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.36)
3% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.36)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.79)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 158
\-------------------------
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7305; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.55)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.25)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.32)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.29)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.69)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.11)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.34)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.37)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.48)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.81)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.89)
7% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 159
\-------------------------
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7291; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.15)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.65)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.75)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.78)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.49)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.30)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded -0.30)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.06)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.85)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.28)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.46)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 160
\-------------------------
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7276; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.26)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.04)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.03)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.26)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.01)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.29)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.80)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 161
\-------------------------
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7261; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.58)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.69)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.61)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.58)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 162
\-------------------------
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7247; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -11.00)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.23)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.47)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.99)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.61)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 163
\-------------------------
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7233; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.84)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.79)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.71)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.36)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.08)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.24)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.19)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.39)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 164
\-------------------------
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7218; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.49)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.14)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.07)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.51)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.40)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.37)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.73)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.96)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.28)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.81)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded -0.34)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.10)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.46)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.46)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.71)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 165
\-------------------------
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7204; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.83)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.50)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.75)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.45)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 166
\-------------------------
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7189; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.65)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.71)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.84)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.06)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 167
\-------------------------
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7175; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.62)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.55)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.12)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.02)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 168
\-------------------------
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7161; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.98)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.63)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.01)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.17)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.50)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.38)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.31)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.46)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.68)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.13)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.85)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.45)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.38)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 1.43)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.21)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.99)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.66)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.55)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.82)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 169
\-------------------------
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7146; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.36)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.62)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.95)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.98)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.70)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.31)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 0.63)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.86)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.62)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent drove forward instead of right. (rewarded -0.10)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.93)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 0.39)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 1.74)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.09)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 0.65)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 170
\-------------------------
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7132; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.16)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.20)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.06)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.71)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.31)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.48)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.20)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.25)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.56)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 171
\-------------------------
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7118; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.77)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.54)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.01)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.11)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.43)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.43)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.02)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.66)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 172
\-------------------------
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7103; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.01)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.94)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.53)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.81)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.27)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.37)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent drove forward instead of left. (rewarded -0.37)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.83)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 173
\-------------------------
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7089; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.44)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.34)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.31)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 174
\-------------------------
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7075; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.44)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.44)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.25)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.78)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.58)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.77)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.11)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.10)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.80)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.80)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.55)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded -0.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 175
\-------------------------
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7061; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.06)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.46)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.20)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.10)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.09)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.70)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.56)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.62)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.03)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.42)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 176
\-------------------------
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7047; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.74)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.86)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.04)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.70)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.24)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 177
\-------------------------
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7033; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.83)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.03)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.68)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.90)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.79)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.08)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.40)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.24)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.41)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'right')
Agent drove right instead of left. (rewarded -0.21)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.66)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.50)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.55)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 178
\-------------------------
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7019; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.27)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.58)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.79)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.14)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.97)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 179
\-------------------------
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.7005; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 1.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.58)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.71)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.97)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.67)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.74)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.74)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.88)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.60)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.30)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.02)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.43)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.95)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 180
\-------------------------
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6991; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.68)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.13)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.45)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.45)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent drove right instead of left. (rewarded 0.30)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.90)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.92)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.97)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.58)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.65)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.07)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.18)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.33)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 181
\-------------------------
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6977; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.92)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.79)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.24)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.54)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.59)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 182
\-------------------------
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6963; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.58)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.92)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.80)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.25)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.93)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.75)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.46)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.73)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.03)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.24)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.14)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.80)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.56)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.36)
3% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 183
\-------------------------
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6949; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.14)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.07)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.15)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.24)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 184
\-------------------------
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6935; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.68)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.86)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.77)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.65)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.96)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.40)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 185
\-------------------------
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6921; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.60)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.81)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.46)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.50)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.02)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.28)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.95)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.14)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 186
\-------------------------
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6907; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.25)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.32)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.50)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.09)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.67)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.93)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.09)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.32)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.83)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.77)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.56)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.38)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.53)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 187
\-------------------------
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6894; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.78)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.13)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.95)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.39)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.42)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.89)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 0.95)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.22)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.86)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.09)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.38)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.30)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 188
\-------------------------
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6880; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.07)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.29)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.05)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.24)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.08)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.98)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.28)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.26)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.81)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.62)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.41)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.73)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.32)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.40)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 0.27)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 189
\-------------------------
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6866; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.67)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent drove forward instead of right. (rewarded 1.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.16)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.81)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.51)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.05)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.59)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.91)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.19)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.86)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.95)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.83)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 190
\-------------------------
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6852; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.69)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.97)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.18)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.18)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 191
\-------------------------
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6839; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.10)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.95)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.57)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.62)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.75)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.16)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 192
\-------------------------
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6825; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.50)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.65)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.80)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.58)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 193
\-------------------------
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6811; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.66)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.62)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.33)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.16)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.27)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.07)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 194
\-------------------------
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6798; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.06)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.47)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.37)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.93)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.35)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.10)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.50)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.73)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.83)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 195
\-------------------------
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6784; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.53)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.81)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 0.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.85)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.46)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.66)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.16)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.59)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'right')
Agent followed the waypoint right. (rewarded 2.08)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.70)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.72)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.40)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.53)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.31)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.88)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 196
\-------------------------
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6771; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.16)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 1.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.21)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.97)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.43)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.17)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.69)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.78)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.77)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.22)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 197
\-------------------------
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6757; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.40)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.79)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.97)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.28)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.15)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.20)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.49)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.97)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.88)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 198
\-------------------------
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6744; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.02)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.08)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.75)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.49)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.68)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.76)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.12)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.24)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.75)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.41)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.38)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.80)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.24)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 199
\-------------------------
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6730; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.37)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.41)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.54)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.36)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.43)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 200
\-------------------------
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6717; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.75)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.58)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.22)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.50)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.93)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.48)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
54% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.38)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.01)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.37)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.07)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.44)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
6% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.60)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.67)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.62)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 201
\-------------------------
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6703; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.26)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.04)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.38)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.75)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.75)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.47)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.86)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.06)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.06)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 202
\-------------------------
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6690; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.53)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.48)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.07)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.96)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.20)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.73)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.54)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 2.05)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.24)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.68)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.00)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.57)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.03)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 203
\-------------------------
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6676; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.98)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.49)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.77)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.64)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.27)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.47)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.29)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.28)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 204
\-------------------------
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6663; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove left instead of right. (rewarded 1.19)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.37)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.30)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.01)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.86)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.76)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
53% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 205
\-------------------------
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6650; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.92)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.78)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.97)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.31)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.02)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.25)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.73)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.27)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.47)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.54)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 206
\-------------------------
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6637; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.13)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.15)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.59)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.45)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.11)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.65)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.21)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded -0.17)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.77)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.75)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.12)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.06)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.33)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 207
\-------------------------
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6623; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.10)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -11.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.60)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.02)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.48)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 1.61)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.38)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.52)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.37)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 208
\-------------------------
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6610; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded 1.26)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.92)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.46)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.79)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.50)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.11)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.53)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.91)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.53)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.03)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.87)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.66)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.36)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 209
\-------------------------
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6597; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.97)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.21)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.86)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.96)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.91)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.43)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.63)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.81)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.59)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.56)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.27)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.83)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.94)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.67)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 210
\-------------------------
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6584; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.46)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.61)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 2.51)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 211
\-------------------------
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6570; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.08)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.71)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.30)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 1.52)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 2.00)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 212
\-------------------------
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6557; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.78)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.43)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.73)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.68)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.76)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.82)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.31)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.01)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.60)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.45)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.44)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.73)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded -0.35)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.44)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 213
\-------------------------
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6544; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.20)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.66)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 214
\-------------------------
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6531; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -10.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.20)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.57)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.78)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.29)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.04)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.62)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.81)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.79)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.54)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 215
\-------------------------
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6518; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.00)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.48)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.80)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.24)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 216
\-------------------------
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6505; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.20)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.68)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.87)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.25)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.40)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.50)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.63)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.70)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.06)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.16)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.19)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.94)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.00)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 217
\-------------------------
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6492; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 0.80)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.73)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.13)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.44)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded 1.32)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.88)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.66)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.33)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove forward instead of right. (rewarded 0.01)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.07)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.41)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded -0.32)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded -0.32)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.91)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.06)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 218
\-------------------------
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6479; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.32)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.30)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.11)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.83)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.81)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.74)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded -0.23)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.99)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.65)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 219
\-------------------------
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6466; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.69)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.81)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.05)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.17)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.90)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.23)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.78)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.30)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.73)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.69)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.04)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.86)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 220
\-------------------------
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6453; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.28)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.36)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.87)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.76)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 221
\-------------------------
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6440; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.33)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.35)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.27)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.64)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 222
\-------------------------
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6427; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.17)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'right')
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.62)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.19)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.23)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.92)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.11)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.37)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 223
\-------------------------
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6415; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.44)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.10)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 1.89)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.29)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.72)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.99)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.38)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent drove forward instead of left. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.41)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.01)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.90)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.13)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 224
\-------------------------
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6402; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.72)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.41)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.25)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.35)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.37)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.45)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.82)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.45)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.15)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent drove forward instead of left. (rewarded 0.74)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.15)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.69)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.29)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.87)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.27)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.59)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded -0.05)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 225
\-------------------------
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6389; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.61)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.88)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.18)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.30)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.36)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 226
\-------------------------
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6376; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 0.11)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.55)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.65)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.53)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.50)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.69)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.46)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.91)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.07)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.36)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.45)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.06)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded -0.10)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 227
\-------------------------
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6364; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.31)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.00)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.58)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.58)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.85)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.93)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.59)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.46)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.01)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.78)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.24)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.94)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 228
\-------------------------
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6351; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.54)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.39)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.23)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.21)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.05)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.88)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.65)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 0.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.95)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.91)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.92)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 229
\-------------------------
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6338; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.33)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.90)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.43)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.71)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.08)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.96)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.99)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 230
\-------------------------
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6325; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.39)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.29)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.82)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.55)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 1.43)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.13)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.29)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.34)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded -0.05)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.53)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.38)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.81)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.12)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.66)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 231
\-------------------------
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6313; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 1.84)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.79)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.73)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.70)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.46)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.65)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.15)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.62)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.69)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.18)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 0.04)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.26)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.57)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.68)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.56)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 232
\-------------------------
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6300; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.32)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.22)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.05)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 1.94)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.34)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.86)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.88)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.76)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.51)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.74)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.32)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
51% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.26)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.71)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
43% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.22)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.68)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.80)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 2.44)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.38)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.21)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.10)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
6% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.77)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
3% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.19)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 233
\-------------------------
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6288; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.25)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.86)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.35)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.03)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.79)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.79)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.31)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.71)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.49)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.11)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.15)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 234
\-------------------------
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6275; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove forward instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.50)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.89)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.41)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.88)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.48)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.94)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.57)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.18)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.72)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.16)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.90)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.62)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 235
\-------------------------
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6263; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.60)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.33)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.38)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.05)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 236
\-------------------------
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6250; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.55)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded 0.52)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.46)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.59)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.03)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.23)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 237
\-------------------------
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6238; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.17)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.91)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.98)
76% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 238
\-------------------------
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6225; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.96)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.95)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.64)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent followed the waypoint forward. (rewarded 1.86)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.06)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.88)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 239
\-------------------------
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6213; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.77)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.96)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.61)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.57)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.51)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.70)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.32)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.56)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.00)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.40)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent drove right instead of left. (rewarded 0.75)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 240
\-------------------------
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6200; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.88)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.65)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.57)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.76)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.14)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.49)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.62)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.21)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.69)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.92)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.28)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded -0.11)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 241
\-------------------------
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6188; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.08)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.84)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.32)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.63)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 0.54)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.89)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.22)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.12)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.48)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded -0.37)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.61)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.42)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 242
\-------------------------
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6175; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent drove right instead of left. (rewarded 0.17)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.50)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.98)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.96)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.01)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.51)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.61)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.40)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.05)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.99)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.40)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.16)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 243
\-------------------------
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6163; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.67)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.75)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.58)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.74)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.77)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.31)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.77)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.20)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -21.00)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.28)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.92)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.58)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.88)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.55)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.85)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.08)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.72)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 0.72)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.57)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.96)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 244
\-------------------------
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6151; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.55)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.91)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.44)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.94)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.50)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.38)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.06)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.83)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.06)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded -0.17)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.29)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 245
\-------------------------
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6139; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.36)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.05)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.16)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.55)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 246
\-------------------------
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6126; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.08)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.04)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.45)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.76)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 247
\-------------------------
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6114; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.10)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.63)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 0.70)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.38)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.27)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.76)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 248
\-------------------------
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6102; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.32)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.66)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.45)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.04)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.14)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.84)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove forward instead of left. (rewarded -0.23)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove forward instead of left. (rewarded -0.23)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.62)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.21)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.10)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.01)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.67)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.65)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 249
\-------------------------
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6090; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.12)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.93)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.72)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.71)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.80)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.71)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.65)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.23)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.54)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 0.04)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.70)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 250
\-------------------------
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6077; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.90)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.66)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.36)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 251
\-------------------------
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6065; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.47)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.89)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.02)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.90)
67% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 252
\-------------------------
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6053; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.28)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.14)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded -0.00)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.24)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove forward instead of left. (rewarded 1.02)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.08)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.39)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.54)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.34)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.53)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.71)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.08)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.56)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 253
\-------------------------
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6041; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.18)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.20)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.14)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.38)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.46)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.37)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.40)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.65)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.25)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 254
\-------------------------
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6029; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.01)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.06)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 255
\-------------------------
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6017; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.61)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.66)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.09)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.58)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.52)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded -0.12)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.76)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.90)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.38)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.51)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.81)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 256
\-------------------------
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
Simulating trial. . .
epsilon = 0.6005; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.47)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.40)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.15)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.18)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.43)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.25)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.98)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.39)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.96)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 257
\-------------------------
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5993; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.11)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.68)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.79)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 258
\-------------------------
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5981; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.37)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.46)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded -0.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded -0.18)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.79)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.34)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.56)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.82)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.63)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded -0.04)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 259
\-------------------------
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5969; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.65)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.31)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.22)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.54)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.11)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.41)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.20)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.20)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 0.48)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.57)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.93)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.54)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.72)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.30)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 260
\-------------------------
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5957; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.14)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.39)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.79)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.51)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.09)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 261
\-------------------------
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5945; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.99)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.74)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.80)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 262
\-------------------------
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5933; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.79)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.78)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.82)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.26)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.89)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.61)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.21)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.12)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.18)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.17)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded -0.17)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.75)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.22)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 263
\-------------------------
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5921; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 1.84)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.70)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 1.75)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.11)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.73)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.34)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.02)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.46)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.00)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.17)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 264
\-------------------------
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5910; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.18)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.22)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.82)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.89)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent drove forward instead of left. (rewarded -0.04)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.64)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.54)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.55)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 265
\-------------------------
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5898; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.84)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.82)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.72)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.91)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.56)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.24)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.94)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.75)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.71)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.10)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 266
\-------------------------
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5886; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.04)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.70)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.51)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 1.17)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.20)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.81)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.10)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.20)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.62)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.72)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.72)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.50)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.59)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.34)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.45)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.76)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.28)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.29)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.29)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.01)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.09)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.96)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.56)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 267
\-------------------------
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5874; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.20)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.49)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.73)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.63)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.71)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.40)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.01)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.20)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.20)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.39)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 268
\-------------------------
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5863; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.87)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.63)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 269
\-------------------------
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5851; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.10)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.96)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.25)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.26)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 0.34)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.13)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.60)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.28)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.62)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.57)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.30)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.44)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 270
\-------------------------
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5839; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.28)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.32)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.76)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.90)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.51)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.19)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 271
\-------------------------
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5827; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.51)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.50)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.66)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.04)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.29)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.24)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.97)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.20)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.32)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 272
\-------------------------
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5816; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.61)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.41)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.71)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.05)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.60)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.91)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.16)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.17)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.17)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.94)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.21)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.89)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.79)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
29% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 273
\-------------------------
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5804; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.40)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.01)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.80)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.94)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.38)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 0.85)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.16)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 274
\-------------------------
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5793; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.67)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.19)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.19)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.01)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.18)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.86)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.88)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.64)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 275
\-------------------------
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5781; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.49)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.60)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.87)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 276
\-------------------------
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5769; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.48)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.41)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.03)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.28)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.10)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 277
\-------------------------
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5758; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.50)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.25)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.71)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.71)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.65)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.65)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 1.68)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.84)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 278
\-------------------------
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5746; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove forward instead of left. (rewarded 1.88)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.23)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.15)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.82)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.04)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.87)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.58)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 279
\-------------------------
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5735; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.34)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.66)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.80)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.13)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.11)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.29)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.67)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.13)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.41)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.33)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.33)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
46% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.35)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.87)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.13)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.87)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
26% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 280
\-------------------------
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5724; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.79)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.18)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.60)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.71)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 281
\-------------------------
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5712; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.17)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.56)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.75)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.33)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.75)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.80)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.39)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent drove right instead of forward. (rewarded -0.03)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.84)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.11)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.80)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 282
\-------------------------
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5701; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.56)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.79)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.77)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.40)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.92)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.21)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.49)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.49)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.70)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.84)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.47)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.46)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 283
\-------------------------
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5689; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.97)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.80)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.10)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.14)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.93)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.87)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.15)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.78)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.38)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.38)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.03)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.01)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.27)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 284
\-------------------------
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5678; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.26)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.46)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.50)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.01)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.01)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.52)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.64)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.72)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.72)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.31)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.78)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 285
\-------------------------
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5667; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.71)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.81)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.68)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.79)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.41)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.95)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.03)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.52)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.77)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.44)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.61)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.60)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.04)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.47)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 286
\-------------------------
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5655; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.48)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.90)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.83)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.23)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.71)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.49)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 287
\-------------------------
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5644; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.18)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.49)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.26)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.38)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.30)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.02)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.08)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.48)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.87)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.21)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
46% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.32)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.86)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.62)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.46)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.05)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.69)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.34)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.32)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.29)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.13)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.74)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.28)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 288
\-------------------------
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5633; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded 0.10)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.83)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 1.32)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.49)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.64)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 289
\-------------------------
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5621; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.00)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.55)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.27)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 290
\-------------------------
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5610; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.78)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.78)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.63)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.77)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.31)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.71)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.76)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.06)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 291
\-------------------------
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5599; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 0.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.07)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.05)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.82)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 292
\-------------------------
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5588; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.49)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.51)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.23)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.69)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.88)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.21)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.03)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.30)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.92)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.44)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 293
\-------------------------
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5577; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.20)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.02)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.76)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.80)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.74)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.16)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove right instead of left. (rewarded 1.38)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.02)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.86)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.53)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.94)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded -0.29)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.20)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.41)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 294
\-------------------------
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5565; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.88)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.13)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.29)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.68)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 295
\-------------------------
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5554; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.80)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.87)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.67)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.39)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.07)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.67)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.74)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.74)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.43)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.67)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.33)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 0.64)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.97)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.84)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.63)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 296
\-------------------------
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5543; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.15)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.37)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.63)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.68)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.74)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.19)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded -0.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.21)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 297
\-------------------------
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5532; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent drove right instead of left. (rewarded 0.79)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.39)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.93)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.97)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.11)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.85)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.04)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.40)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.60)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.20)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.15)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.57)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 298
\-------------------------
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5521; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.63)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.09)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.04)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.82)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.49)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 299
\-------------------------
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5510; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.31)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.46)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.19)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.61)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.66)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.83)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.78)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.23)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.18)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.61)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.37)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.04)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.37)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.24)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.61)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.70)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.47)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.47)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded -0.01)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.01)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.87)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded -0.18)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.41)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.63)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.17)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 300
\-------------------------
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5499; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.70)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.44)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.04)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.85)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 301
\-------------------------
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5488; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.82)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.60)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.34)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.06)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 302
\-------------------------
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5477; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.26)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 0.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.13)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.08)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.24)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.51)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.77)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.30)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.05)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.83)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.46)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.69)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.30)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.30)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.60)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.30)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.00)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 303
\-------------------------
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5466; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.67)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.34)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.12)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.82)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.33)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.08)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.50)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.87)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.41)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.38)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.02)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.81)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.09)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.13)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 304
\-------------------------
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5455; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.17)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.14)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.68)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.10)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.59)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 305
\-------------------------
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5444; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.98)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.92)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.39)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.96)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded -0.19)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.03)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.97)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 306
\-------------------------
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5434; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.49)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -9.97)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.00)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.16)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.84)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.14)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.45)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.96)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.20)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 307
\-------------------------
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5423; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.01)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.03)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.10)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.10)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.23)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.84)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.30)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.46)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 308
\-------------------------
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5412; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.93)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.59)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.01)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.07)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.90)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.02)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 309
\-------------------------
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5401; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.56)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.95)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.63)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.42)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.49)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.50)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.49)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.58)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded -0.25)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.82)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.04)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.37)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.03)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.27)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 310
\-------------------------
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5390; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.26)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.37)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.15)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.28)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.92)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.59)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 311
\-------------------------
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5379; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.74)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.20)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.39)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 312
\-------------------------
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5369; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.37)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.36)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.12)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.47)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.66)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.78)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.68)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.96)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.80)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.39)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.66)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.38)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.79)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.90)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.27)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 313
\-------------------------
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5358; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.58)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.43)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.88)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.28)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.32)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.15)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.48)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.57)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.59)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 314
\-------------------------
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5347; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.87)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.32)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.86)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.07)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.14)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.05)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.77)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.10)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 0.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 315
\-------------------------
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5337; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.46)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.41)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.11)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.37)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.53)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.65)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.08)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.49)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 316
\-------------------------
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5326; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.56)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.90)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.40)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.02)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.95)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.64)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.64)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.71)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.71)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.23)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.65)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.07)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.86)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 0.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 317
\-------------------------
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5315; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.81)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.45)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.38)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.48)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.78)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.00)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.11)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.70)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.66)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 318
\-------------------------
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5305; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.17)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.68)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.67)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded -0.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.40)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded -0.22)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.93)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.57)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.24)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.19)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.71)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.71)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.09)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.44)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.59)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 319
\-------------------------
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5294; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.21)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.84)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.85)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded -0.58)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 320
\-------------------------
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5283; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.07)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.37)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.90)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.59)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.30)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.63)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.61)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 321
\-------------------------
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5273; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.10)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.82)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.71)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.62)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.47)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.50)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.04)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.86)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.16)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.25)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded -0.11)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 322
\-------------------------
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5262; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.84)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.95)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.17)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.56)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.33)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.78)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.98)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.47)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.20)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.01)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.61)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
33% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.81)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.31)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.03)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.23)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 0.24)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 0.66)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.09)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.33)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.33)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.68)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.29)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 323
\-------------------------
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5252; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.26)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.24)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.61)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.84)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.08)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
67% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 324
\-------------------------
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5241; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.52)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.28)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 1.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 325
\-------------------------
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5231; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.29)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.96)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.27)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.22)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 1.70)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.48)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.28)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.93)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 326
\-------------------------
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5220; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.00)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.27)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.16)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.79)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.60)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 327
\-------------------------
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5210; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.52)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.64)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.37)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.57)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded -0.25)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.61)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.27)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.73)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded -0.48)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 328
\-------------------------
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5200; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.20)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.84)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.19)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.19)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.43)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.32)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.06)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.36)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.18)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.40)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.88)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.05)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 329
\-------------------------
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5189; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.74)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.99)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.60)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.56)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.46)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.12)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.07)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.98)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.99)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.80)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.94)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.29)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.68)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 330
\-------------------------
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5179; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -9.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.78)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.95)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.08)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.58)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.11)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.06)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.48)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.68)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.10)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.53)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.64)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 331
\-------------------------
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5169; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.15)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.44)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent drove right instead of left. (rewarded 1.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 1.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.17)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.07)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 332
\-------------------------
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5158; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.72)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.90)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.02)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.15)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.23)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.95)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.11)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.28)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.60)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.45)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.45)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.97)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.79)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.10)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.67)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 333
\-------------------------
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5148; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.69)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 334
\-------------------------
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5138; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.24)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.35)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.60)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.91)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.91)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.50)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.18)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.67)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.29)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.74)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.56)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'right')
Agent followed the waypoint left. (rewarded 2.41)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.45)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 1.51)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.67)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.45)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.63)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.49)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.79)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.38)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.79)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
7% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 335
\-------------------------
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5127; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.90)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.08)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.72)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.27)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.33)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.77)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.21)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.79)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.88)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.59)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 0.43)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent drove forward instead of right. (rewarded 0.13)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded -0.68)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.78)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 336
\-------------------------
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5117; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded 0.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.12)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.33)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.33)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.05)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove forward instead of left. (rewarded 0.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.10)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.55)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.08)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.62)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.23)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded -0.58)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.64)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.04)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 337
\-------------------------
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5107; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.63)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.17)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.18)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.98)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 0.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.88)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.23)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 338
\-------------------------
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5097; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.47)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.38)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.80)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.40)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.40)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.96)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.75)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.03)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.65)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.05)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.75)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 339
\-------------------------
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5086; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.70)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.35)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.85)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.25)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.73)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.15)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 340
\-------------------------
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5076; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.18)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.77)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.92)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.22)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.20)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.56)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.36)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.23)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.50)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 341
\-------------------------
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5066; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.36)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.50)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.33)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.69)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded -0.07)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.65)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.62)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.79)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.84)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.60)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.60)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove right instead of left. (rewarded -0.48)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 342
\-------------------------
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5056; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.00)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.24)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.86)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.18)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.80)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 343
\-------------------------
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5046; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.26)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.58)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.09)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.22)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.71)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.03)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.96)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 344
\-------------------------
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5036; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.42)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.78)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.34)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.07)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.69)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.62)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.87)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.52)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.98)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.21)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.32)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.85)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded -0.24)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 345
\-------------------------
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5026; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.00)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.57)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.96)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.62)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.41)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.62)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 346
\-------------------------
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5016; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.64)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.63)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.89)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.55)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.17)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 0.96)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.12)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.52)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.62)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.43)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.09)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.77)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.05)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded -0.53)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.09)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.22)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 347
\-------------------------
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.5006; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.37)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.14)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 348
\-------------------------
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4996; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.42)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.01)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.89)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.44)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.44)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.68)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.57)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.14)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.10)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.40)
53% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 349
\-------------------------
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4986; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.13)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.48)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.91)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.77)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 350
\-------------------------
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4976; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.81)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.33)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove left instead of right. (rewarded -0.17)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.77)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.65)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.56)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.16)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.19)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.99)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.11)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.47)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 351
\-------------------------
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4966; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.50)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 1.97)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.58)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.29)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.04)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.98)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.80)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.18)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.29)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.81)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 352
\-------------------------
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4956; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.64)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.13)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.65)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.05)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.45)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded -0.56)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 353
\-------------------------
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4946; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'right')
Agent followed the waypoint right. (rewarded 2.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.92)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 354
\-------------------------
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4936; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.32)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.90)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.92)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.58)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.93)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 355
\-------------------------
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4926; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.72)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.55)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 356
\-------------------------
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4916; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.20)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.58)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.19)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.16)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 0.51)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.64)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.33)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.93)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.05)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 357
\-------------------------
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4907; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.80)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.62)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.50)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.80)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.63)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.12)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.19)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 0.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 358
\-------------------------
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4897; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.59)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.43)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.05)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.60)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.70)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.36)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 359
\-------------------------
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4887; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.24)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.61)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.00)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.63)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.19)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.58)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.80)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.11)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.63)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.12)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.79)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.32)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.30)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.10)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.06)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.27)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.70)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.24)
7% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 360
\-------------------------
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4877; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove left instead of forward. (rewarded 1.56)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.38)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.52)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.45)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.41)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.66)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.78)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.20)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.50)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.67)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.92)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.28)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.68)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.88)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.20)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.34)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded -0.15)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.93)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.99)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.72)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 361
\-------------------------
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4868; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.77)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent drove left instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.77)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.79)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.85)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.55)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.34)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.25)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.13)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.86)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 362
\-------------------------
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4858; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.32)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.50)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.32)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.54)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.94)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 363
\-------------------------
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4848; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.77)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.27)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.45)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.77)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.75)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.72)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.34)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.34)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.12)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.13)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.40)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 364
\-------------------------
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4838; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.68)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.60)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.20)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.18)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.18)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.03)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.32)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.29)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.30)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.51)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.11)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.28)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.28)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.87)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.31)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'right')
Agent drove right instead of left. (rewarded -0.19)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.13)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 365
\-------------------------
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4829; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.03)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.51)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.06)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.83)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.37)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.79)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.15)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.95)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 366
\-------------------------
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4819; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.10)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove forward instead of right. (rewarded 1.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 367
\-------------------------
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4809; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 1.68)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.41)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.71)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.41)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.47)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.24)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.92)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.48)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.12)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.04)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.63)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.78)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.13)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 368
\-------------------------
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4800; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.11)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.72)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.50)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded 1.05)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.67)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.61)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.83)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.89)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.91)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.78)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.57)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.79)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded -0.19)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.80)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.42)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.21)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.43)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.72)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 369
\-------------------------
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4790; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.71)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.86)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.61)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.98)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.24)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.09)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.49)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 370
\-------------------------
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4781; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.80)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.74)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.70)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.74)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.08)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 371
\-------------------------
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4771; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.95)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.61)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 372
\-------------------------
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4762; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.91)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.87)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.44)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.53)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.45)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.20)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.43)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.00)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.07)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.38)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.25)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.48)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.48)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 0.61)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 373
\-------------------------
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4752; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.69)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.92)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.83)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 374
\-------------------------
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4743; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.63)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.28)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.30)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.05)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.25)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.38)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.63)
37% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 375
\-------------------------
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4733; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.61)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded -0.01)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.83)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove forward instead of right. (rewarded -0.19)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.36)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove forward instead of right. (rewarded 1.06)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.20)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 376
\-------------------------
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4724; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.56)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.33)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.60)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.07)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.09)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.24)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.28)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.38)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 377
\-------------------------
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4714; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.04)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.12)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.99)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent drove right instead of forward. (rewarded -0.15)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.69)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.88)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 378
\-------------------------
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4705; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.44)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.23)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.77)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.70)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.74)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 379
\-------------------------
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4695; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.62)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.76)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.00)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.89)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.56)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.81)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.51)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.32)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.66)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.61)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 380
\-------------------------
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4686; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.41)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.74)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.23)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.95)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.87)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.79)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.14)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.98)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.20)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 381
\-------------------------
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4677; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.22)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.70)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.14)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.45)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.84)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.17)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded 0.27)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.68)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.35)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.85)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.51)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.16)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.57)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.10)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent drove right instead of forward. (rewarded -0.13)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 382
\-------------------------
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4667; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.21)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.27)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 1.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.35)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.40)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.49)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.14)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 383
\-------------------------
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4658; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 2.21)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.95)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.49)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.75)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.55)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.91)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded 1.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.20)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove forward instead of right. (rewarded 0.12)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 384
\-------------------------
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4649; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.29)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.28)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.77)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.37)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.37)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.89)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.62)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 385
\-------------------------
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4639; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.78)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.92)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.30)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.77)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.75)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.27)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.18)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.46)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.46)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.06)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.61)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.75)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.01)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.38)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.11)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.04)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.24)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.37)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.72)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.04)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 386
\-------------------------
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4630; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.66)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.69)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.60)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.58)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.62)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.44)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.53)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.15)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.24)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.98)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.41)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.58)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.46)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.26)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.36)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 387
\-------------------------
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4621; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.01)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.33)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.30)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.10)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.79)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.22)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.73)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.73)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.51)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.10)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.80)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.49)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.90)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.91)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.17)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.29)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.71)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 388
\-------------------------
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4612; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.25)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.61)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 0.71)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.15)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.15)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.31)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.16)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 0.94)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 389
\-------------------------
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4602; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.49)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.24)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.93)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.17)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.47)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.11)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.77)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.65)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.06)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.07)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.92)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.62)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 390
\-------------------------
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4593; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.82)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 1.83)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.66)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.36)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.56)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.51)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.12)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.96)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.73)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.10)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 1.12)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.67)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 391
\-------------------------
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4584; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.34)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.84)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.61)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.48)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.00)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.12)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.42)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.40)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.10)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 392
\-------------------------
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4575; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent drove forward instead of right. (rewarded 1.15)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.57)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.90)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.12)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.76)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.71)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 393
\-------------------------
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4566; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.35)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.78)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.45)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.71)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.92)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.61)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.31)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.04)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.04)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.48)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.48)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.16)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 394
\-------------------------
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4557; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.75)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.39)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.32)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.16)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 395
\-------------------------
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4548; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.32)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.01)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.38)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.37)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 396
\-------------------------
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4538; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.37)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.01)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 1.57)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.80)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.59)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded -0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 397
\-------------------------
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4529; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.51)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.13)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.60)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.92)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.24)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 1.84)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.78)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 0.83)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.31)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.72)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.29)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.84)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.50)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.62)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.28)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.54)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 398
\-------------------------
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4520; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.27)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.37)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.61)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.08)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.59)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.11)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 399
\-------------------------
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4511; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.32)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.22)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 1.48)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.21)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.14)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.71)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.69)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.17)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.30)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.41)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.36)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.15)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 400
\-------------------------
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4502; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.80)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.57)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.15)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.33)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.17)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.55)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 401
\-------------------------
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4493; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove left instead of right. (rewarded 1.10)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
90% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.54)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.98)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.06)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.94)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.50)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.70)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 2.13)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.99)
53% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 402
\-------------------------
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4484; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 0.36)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.63)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.47)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.29)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.78)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 403
\-------------------------
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4475; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.89)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.67)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.73)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.70)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.09)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.15)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.78)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.41)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.62)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.67)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.00)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.73)
47% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 404
\-------------------------
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4466; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.12)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove forward instead of left. (rewarded 0.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.48)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.70)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.44)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove forward instead of right. (rewarded 0.02)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.49)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.54)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.05)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.07)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.58)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.22)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.59)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.42)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded -0.25)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 405
\-------------------------
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4457; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.51)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.33)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.91)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.35)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.30)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.47)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 1.53)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.75)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.57)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.70)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.82)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.84)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.08)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 406
\-------------------------
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4449; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.58)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.07)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.72)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.88)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.59)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.68)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.64)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.70)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.35)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.06)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.01)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove right instead of left. (rewarded -0.02)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.64)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.62)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 407
\-------------------------
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4440; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.09)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.20)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 408
\-------------------------
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4431; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.15)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.65)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.51)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.41)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'right')
Agent followed the waypoint left. (rewarded 2.65)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.96)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.03)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.21)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.10)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.34)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.34)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.10)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.93)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.59)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 409
\-------------------------
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4422; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent drove right instead of forward. (rewarded 0.45)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.07)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.57)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.24)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.84)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.00)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.78)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.48)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.73)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 410
\-------------------------
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4413; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.81)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.91)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 2.21)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.02)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 0.22)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.58)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.48)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.69)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.74)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.36)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 411
\-------------------------
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4404; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.01)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.46)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.69)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.98)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.41)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.06)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.65)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 1.65)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.95)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.25)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.19)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.50)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.72)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.07)
49% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 412
\-------------------------
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4396; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.74)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.19)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.60)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 413
\-------------------------
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4387; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.78)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.75)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.59)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.06)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.75)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 1.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.43)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 414
\-------------------------
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4378; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.75)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.77)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.51)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.19)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.91)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.57)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.87)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.00)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.26)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.89)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.65)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 415
\-------------------------
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4369; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.30)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 0.06)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.35)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.35)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.90)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.99)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.36)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.58)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.17)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 416
\-------------------------
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4360; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.34)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.11)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.71)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.04)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.34)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.36)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.28)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove forward instead of left. (rewarded 1.28)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded -0.08)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.13)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 0.46)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent drove left instead of forward. (rewarded 0.86)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 417
\-------------------------
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4352; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.33)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.56)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.37)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 418
\-------------------------
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4343; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 1.64)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.22)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.98)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.48)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.75)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.10)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.13)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.40)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.46)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.69)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.35)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.44)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 419
\-------------------------
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4334; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.38)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.07)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.78)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.38)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.68)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 420
\-------------------------
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4326; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.31)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.85)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.92)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 0.92)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 421
\-------------------------
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4317; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.04)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.39)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.21)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.86)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.36)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 422
\-------------------------
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4308; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.24)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.06)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.00)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.90)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 423
\-------------------------
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4300; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 0.99)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.50)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.42)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.07)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent drove right instead of forward. (rewarded 1.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.96)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 424
\-------------------------
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4291; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.78)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.05)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.10)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.54)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 1.37)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.41)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.62)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.53)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.91)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.73)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.13)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.31)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.56)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.46)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.67)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.15)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.10)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 425
\-------------------------
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4283; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 1.75)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.02)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.62)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.15)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 426
\-------------------------
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4274; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.07)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.17)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.41)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.34)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.99)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.34)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.73)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.36)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.47)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.02)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.66)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 427
\-------------------------
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4266; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.90)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.85)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.99)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
74% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.59)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.88)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.20)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove right instead of left. (rewarded 1.29)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.53)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.72)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.11)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.13)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.22)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.10)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.82)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.45)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.23)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.77)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.89)
14% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 428
\-------------------------
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4257; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
84% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.32)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.00)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.61)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.61)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 0.21)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.20)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove right instead of left. (rewarded 0.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.06)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.61)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.12)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.12)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.14)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.92)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.48)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.16)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.92)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 429
\-------------------------
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4249; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.90)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.07)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.57)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.32)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.86)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.72)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.84)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.81)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.84)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.90)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 430
\-------------------------
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4240; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.95)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.91)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.01)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.69)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.11)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 431
\-------------------------
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4232; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.97)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.76)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.76)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.63)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.31)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.56)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.10)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded -0.45)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.56)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.74)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 0.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 432
\-------------------------
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4223; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.99)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.87)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.60)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.87)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.13)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.76)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.04)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.04)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.21)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.05)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.12)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded -0.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded 1.10)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -10.88)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 433
\-------------------------
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4215; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.84)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.69)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.53)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.57)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.13)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 0.97)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 434
\-------------------------
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4206; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.01)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.48)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.57)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.68)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.31)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.14)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
63% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 435
\-------------------------
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4198; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.96)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.17)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.34)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.88)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.04)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.58)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 436
\-------------------------
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4190; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.79)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.83)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 437
\-------------------------
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4181; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent properly idled at a red light. (rewarded 2.80)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.26)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.91)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.82)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.92)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.81)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.73)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 0.73)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.64)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded -0.18)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.35)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.03)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.09)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.91)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 438
\-------------------------
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4173; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.98)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.33)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.74)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.55)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 439
\-------------------------
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4164; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 2.09)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.47)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.72)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 440
\-------------------------
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4156; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.49)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.98)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.89)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.42)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.90)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.65)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 441
\-------------------------
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4148; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 0.03)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.79)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.83)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.28)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.97)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove right instead of left. (rewarded 0.65)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.05)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.44)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.17)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.72)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.79)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 442
\-------------------------
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4140; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.56)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.21)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.50)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.92)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.66)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.12)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.95)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 443
\-------------------------
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4131; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.23)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.98)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.15)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.52)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.52)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.54)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.55)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.50)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.91)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.37)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.71)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 444
\-------------------------
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4123; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.74)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.75)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.72)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.34)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.03)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.21)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.65)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.37)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.60)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.91)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.82)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.78)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.78)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.97)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove right instead of left. (rewarded -0.35)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.54)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.75)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.35)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 445
\-------------------------
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4115; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.53)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.58)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.91)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.44)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.51)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.96)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.00)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.56)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.04)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 446
\-------------------------
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4107; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.91)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.13)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.93)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 447
\-------------------------
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4098; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.67)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.49)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.76)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.75)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.75)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.26)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.12)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.55)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.47)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.48)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.41)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.87)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.78)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 448
\-------------------------
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4090; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.92)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.91)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.95)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.55)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.88)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.56)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.53)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.38)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.38)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.93)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.73)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 449
\-------------------------
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4082; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent drove forward instead of right. (rewarded 1.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.40)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.68)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.68)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.00)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.60)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 450
\-------------------------
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4074; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.79)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.05)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.12)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.68)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'left')
Agent drove left instead of forward. (rewarded 1.07)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.67)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.96)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.98)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.72)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
28% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 451
\-------------------------
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4066; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.89)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.14)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.58)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.41)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.65)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 0.81)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.18)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.18)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.05)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.35)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.96)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 452
\-------------------------
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4058; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.71)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.82)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.61)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.89)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.59)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 1.67)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.87)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.94)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.11)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.04)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.46)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.05)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.75)
49% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 453
\-------------------------
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4049; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.69)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.14)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.77)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.05)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.29)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.39)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 454
\-------------------------
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4041; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.93)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.11)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 1.66)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.07)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.37)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.08)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.95)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.79)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.43)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.20)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.23)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.20)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.55)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
33% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 455
\-------------------------
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4033; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.83)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.26)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.02)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.82)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 0.82)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.50)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.27)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.98)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove right instead of left. (rewarded 0.93)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.60)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.73)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.67)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.53)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.93)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 456
\-------------------------
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4025; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.20)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.99)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.99)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.21)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.02)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.26)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.24)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.91)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 457
\-------------------------
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4017; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.52)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 2.24)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.56)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.28)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.67)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.41)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.57)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.51)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.22)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.58)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 458
\-------------------------
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4009; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.50)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.14)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.98)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.49)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.00)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.82)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.51)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.16)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.94)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.67)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.56)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.26)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 459
\-------------------------
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
Simulating trial. . .
epsilon = 0.4001; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.84)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 1.27)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 1.32)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.67)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.92)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.50)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.10)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 1.20)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.28)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.09)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.76)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.99)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.74)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.49)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.93)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.88)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.37)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.23)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.41)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 460
\-------------------------
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3993; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.61)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.74)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.46)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.83)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.33)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.54)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.90)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.29)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.63)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.85)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 461
\-------------------------
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3985; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.74)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.93)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.56)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.82)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 1.12)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent drove left instead of right. (rewarded 0.08)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.16)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.33)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.06)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.98)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.82)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.20)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.12)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.03)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.87)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.46)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.67)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.03)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded -0.21)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
29% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.87)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.08)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.08)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.13)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.12)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 0.44)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.15)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.07)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.36)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 462
\-------------------------
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3977; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.07)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'right')
Agent attempted driving forward through a red light. (rewarded -9.45)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.59)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.76)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.46)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.86)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.97)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.73)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.74)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.61)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.09)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.42)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 463
\-------------------------
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3969; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.19)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.08)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.02)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.17)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.53)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 464
\-------------------------
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3961; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.47)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.63)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 0.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.96)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 465
\-------------------------
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3953; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.97)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.43)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.58)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.05)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.32)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.79)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.65)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.41)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 0.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.16)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.68)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.55)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.10)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.12)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 466
\-------------------------
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3946; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.35)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.74)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.45)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.89)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.51)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.54)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.71)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.84)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.55)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.24)
49% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 467
\-------------------------
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3938; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.30)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.73)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.26)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.96)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.14)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.67)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.93)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.24)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 1.45)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 0.94)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 468
\-------------------------
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3930; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
88% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.24)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.24)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.13)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.51)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.15)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 469
\-------------------------
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3922; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.87)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.00)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.28)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
86% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.96)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.71)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.43)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove left instead of forward. (rewarded 0.72)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.18)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.38)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.57)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.92)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.33)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.71)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.95)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.51)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent drove right instead of left. (rewarded 0.07)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.45)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.58)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.96)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.70)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.65)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 470
\-------------------------
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3914; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.08)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.89)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.07)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.92)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.66)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.25)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.83)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.60)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.86)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 471
\-------------------------
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3906; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.33)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.80)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.12)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.75)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.35)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.28)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.93)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 472
\-------------------------
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3898; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.12)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.85)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.58)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.14)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.18)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.18)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.79)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.00)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded -0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 473
\-------------------------
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3891; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.05)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.66)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.11)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.09)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.99)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.29)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.88)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.12)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.63)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.88)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.23)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.83)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.34)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.87)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 474
\-------------------------
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3883; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.95)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.36)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.39)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.59)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.99)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.87)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.44)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.21)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.59)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.63)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.15)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 475
\-------------------------
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3875; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.61)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.86)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.24)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.92)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.68)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.91)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 476
\-------------------------
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3867; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.72)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.37)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.33)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.15)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.53)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.61)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.88)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.36)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.89)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.59)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.18)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.53)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.53)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.61)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.71)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.65)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.39)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 477
\-------------------------
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3860; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.86)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
91% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.29)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.02)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.68)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.72)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
66% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 478
\-------------------------
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3852; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.68)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.56)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.41)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.50)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.87)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.70)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 479
\-------------------------
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3844; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.24)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.27)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.19)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent drove forward instead of left. (rewarded 0.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.66)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.47)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.01)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 480
\-------------------------
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3837; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.68)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.47)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.63)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 481
\-------------------------
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3829; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.75)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.40)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.52)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.11)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.65)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.89)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.66)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.34)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.35)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.83)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.84)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent drove left instead of right. (rewarded -0.13)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.94)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.33)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.84)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 482
\-------------------------
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3821; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.68)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.47)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.35)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.29)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.53)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.92)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.10)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.07)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 483
\-------------------------
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3814; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.52)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.09)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.85)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.35)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.33)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 484
\-------------------------
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3806; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.38)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.91)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.44)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.10)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.60)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 485
\-------------------------
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3798; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.09)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.91)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.28)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 486
\-------------------------
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3791; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.10)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.24)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.03)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.09)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.88)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.33)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.22)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.66)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.59)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.82)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.05)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 487
\-------------------------
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3783; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.64)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.19)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.85)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.65)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.69)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.32)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.48)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.61)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 488
\-------------------------
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3776; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.78)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.97)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.07)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.89)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.73)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.38)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.07)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.68)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded -0.03)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.03)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.09)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.07)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.24)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.63)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.16)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.63)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.79)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 0.77)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.58)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.40)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 489
\-------------------------
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3768; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.46)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.83)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.23)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.11)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 0.79)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.81)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.51)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.23)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.27)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.26)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.01)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 0.68)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.50)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 490
\-------------------------
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3761; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.41)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.54)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.35)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.29)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.33)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.63)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.71)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.44)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.97)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.03)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.21)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
33% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.21)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.45)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.14)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.14)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.06)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 1.22)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
3% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 491
\-------------------------
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3753; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.73)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.86)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.86)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.81)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.32)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.37)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.72)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.73)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.12)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.20)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 492
\-------------------------
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3746; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.80)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.80)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.63)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.63)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 1.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.63)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.66)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 493
\-------------------------
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3738; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.99)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.29)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.39)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.06)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.87)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.49)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.25)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 494
\-------------------------
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3731; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 2.17)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 0.99)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.50)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.20)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.49)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.14)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded -0.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.33)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.00)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 495
\-------------------------
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3723; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 2.81)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.71)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.80)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.46)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.67)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.44)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 496
\-------------------------
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3716; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.37)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded 0.74)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove forward instead of left. (rewarded 1.00)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.09)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.72)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.72)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.56)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.99)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.11)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.62)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.33)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 497
\-------------------------
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3708; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.97)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent drove forward instead of left. (rewarded 1.49)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.38)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.64)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.10)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.29)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.22)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.06)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.88)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.39)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 498
\-------------------------
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3701; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.94)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.70)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.73)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.27)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.49)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.64)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.32)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.18)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 499
\-------------------------
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3694; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.56)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.69)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.46)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.68)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.04)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.86)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.68)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.07)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.11)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.22)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', 'left')
Agent drove right instead of left. (rewarded -0.28)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.26)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.23)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.94)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 500
\-------------------------
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3686; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.88)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.22)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.16)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.16)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.99)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 501
\-------------------------
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3679; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.39)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 502
\-------------------------
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3671; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.42)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.95)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.08)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.11)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.16)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.27)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.10)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.33)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.22)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.89)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.43)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.43)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 503
\-------------------------
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3664; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.10)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.18)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.95)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.37)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 504
\-------------------------
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3657; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.02)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.77)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.35)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 0.22)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.05)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.31)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.03)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.55)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.99)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.93)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.40)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.62)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 505
\-------------------------
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3649; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.74)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', None)
Agent followed the waypoint right. (rewarded 2.21)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.12)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.12)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.53)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.92)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.35)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded -0.29)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.31)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.94)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 506
\-------------------------
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3642; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.27)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.29)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.95)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.04)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.98)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.33)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.19)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.61)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 507
\-------------------------
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3635; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.80)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.90)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.42)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove forward instead of right. (rewarded 0.51)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 508
\-------------------------
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3628; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.44)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 1.81)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 509
\-------------------------
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3620; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.07)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.45)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded -0.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.03)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.61)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.34)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.29)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 510
\-------------------------
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3613; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.98)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.90)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.45)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 511
\-------------------------
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3606; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.78)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.13)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.23)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.74)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.28)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.28)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 512
\-------------------------
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3599; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.25)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.45)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.04)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'right', 'forward')
Agent drove forward instead of left. (rewarded 1.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.05)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.70)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded -0.30)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.72)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.49)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 513
\-------------------------
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3592; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.23)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
89% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.13)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.42)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.49)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.52)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.13)
71% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.13)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.79)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.99)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.20)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.14)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.74)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.08)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.37)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.72)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 514
\-------------------------
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3584; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.96)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.50)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.14)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.40)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.18)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded -0.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.16)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.09)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.38)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.45)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.45)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.87)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 515
\-------------------------
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3577; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.00)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.69)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.72)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.01)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove right instead of left. (rewarded 0.22)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.39)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.54)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.69)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.58)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.99)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 516
\-------------------------
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3570; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.92)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.62)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.51)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.42)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.59)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.01)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.17)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.99)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.09)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.61)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.80)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.35)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.96)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 517
\-------------------------
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3563; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.49)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.90)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.79)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 518
\-------------------------
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3556; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.04)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.58)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.81)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.20)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent drove left instead of right. (rewarded 0.45)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.19)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.76)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.62)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.42)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.63)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.14)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 519
\-------------------------
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3549; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.33)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.65)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.45)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 520
\-------------------------
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3542; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.06)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.33)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.35)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.64)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.91)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.45)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
77% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 521
\-------------------------
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3535; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.31)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.97)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.74)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.49)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.39)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.92)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 522
\-------------------------
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3527; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.78)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 1.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.24)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.61)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.39)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.56)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.11)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.95)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.93)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent drove right instead of left. (rewarded -0.50)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.05)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.30)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 523
\-------------------------
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3520; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.22)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.60)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent drove right instead of forward. (rewarded 1.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.52)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove left instead of forward. (rewarded 0.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.10)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.35)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.42)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.42)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.86)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.47)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.22)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 0.99)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.23)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.57)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 524
\-------------------------
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3513; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.34)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.79)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.83)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.85)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 525
\-------------------------
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3506; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.16)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.65)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.52)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent followed the waypoint left. (rewarded 1.86)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.03)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 526
\-------------------------
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3499; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.57)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.44)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.93)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.60)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 0.21)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.59)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 0.89)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.23)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 527
\-------------------------
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3492; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.13)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.58)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.05)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.38)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.80)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.83)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.09)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.97)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.50)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.82)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 528
\-------------------------
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3485; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.40)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.36)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.06)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.02)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.56)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.66)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.76)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.39)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.00)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded -0.45)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.86)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.49)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.27)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 529
\-------------------------
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3478; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.34)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 530
\-------------------------
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3471; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.43)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.44)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.06)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.43)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.14)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.35)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.60)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.56)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.03)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.24)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.59)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.29)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 531
\-------------------------
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3465; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded 1.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.19)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.10)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.74)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.20)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.54)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.47)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.13)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.05)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.39)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.05)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.32)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.56)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.89)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.41)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.21)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.62)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.62)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.05)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 532
\-------------------------
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3458; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.50)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.61)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.73)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.50)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.50)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.47)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.29)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.12)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.54)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 533
\-------------------------
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3451; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'right')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.10)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.54)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 2.28)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.59)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.66)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.43)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded -0.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.70)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.87)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
5% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 534
\-------------------------
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3444; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.01)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.31)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.82)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.03)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.30)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.41)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.03)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.51)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove left instead of forward. (rewarded 0.04)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.88)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent drove left instead of forward. (rewarded -0.22)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'right')
Agent drove left instead of right. (rewarded -0.13)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 535
\-------------------------
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3437; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.40)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.99)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 0.98)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.23)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.37)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 536
\-------------------------
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3430; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.22)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.62)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.71)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.64)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.69)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.55)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.19)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.17)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.93)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.65)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.36)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.10)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.34)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 537
\-------------------------
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3423; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.67)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.97)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.27)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.60)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 2.78)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 538
\-------------------------
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3416; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.90)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.83)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.68)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.31)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 0.16)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent attempted driving forward through a red light. (rewarded -9.35)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.83)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.69)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 539
\-------------------------
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3410; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent drove left instead of right. (rewarded 0.42)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'forward')
Agent drove right instead of left. (rewarded 1.27)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.35)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.45)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.69)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 540
\-------------------------
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3403; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 0.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.92)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.55)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.72)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.58)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 541
\-------------------------
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3396; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.31)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.15)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.06)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.41)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent drove left instead of right. (rewarded 0.42)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.93)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.93)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'left')
Agent properly idled at a red light. (rewarded 0.91)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.98)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.51)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 542
\-------------------------
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3389; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.06)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.14)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent drove forward instead of left. (rewarded 0.14)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.62)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.91)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.64)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.19)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.79)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.19)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.19)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.33)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 543
\-------------------------
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3382; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.66)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.84)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.55)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.95)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.83)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 544
\-------------------------
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3376; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.78)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.94)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.62)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.35)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.42)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.50)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.76)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.71)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 545
\-------------------------
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3369; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.49)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.38)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.64)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 546
\-------------------------
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3362; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent drove forward instead of right. (rewarded 1.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.45)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.54)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.97)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 547
\-------------------------
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3355; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.16)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.54)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.01)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.50)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.90)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.54)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 2.15)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 548
\-------------------------
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3349; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.35)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.96)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.38)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.57)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.05)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.43)
28% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 549
\-------------------------
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3342; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.84)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.16)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.34)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 2.78)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.11)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.74)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.82)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 550
\-------------------------
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3335; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.48)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.33)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'left')
Agent drove forward instead of left. (rewarded 0.42)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.73)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 551
\-------------------------
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3329; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.95)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.45)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
84% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.32)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.23)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 1.33)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.38)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.40)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.04)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.78)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.43)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.26)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.69)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.08)
16% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.08)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.31)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.34)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 552
\-------------------------
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3322; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.44)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.93)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.20)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.20)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.28)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.07)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.67)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 553
\-------------------------
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3315; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.64)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.47)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.84)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.91)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 1.54)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.80)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.25)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.51)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
15% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.75)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded -0.54)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.42)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.71)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 554
\-------------------------
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3309; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.63)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.96)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.37)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 555
\-------------------------
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3302; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.37)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.59)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.45)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.90)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 556
\-------------------------
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3296; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.53)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 0.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.64)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.21)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.83)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.58)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 557
\-------------------------
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3289; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.49)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.37)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.64)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.41)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.61)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.10)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent drove forward instead of left. (rewarded 0.68)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.10)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.56)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.79)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.29)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.69)
47% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 558
\-------------------------
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3282; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 559
\-------------------------
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3276; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.96)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.69)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.27)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.94)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.23)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.51)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.82)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.84)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 560
\-------------------------
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3269; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.95)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.41)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.95)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.06)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.14)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.01)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.84)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.80)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.32)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.49)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.28)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.53)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.63)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
16% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.50)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.92)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 561
\-------------------------
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3263; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.94)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.29)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.70)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.69)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.52)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.52)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.81)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.84)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.22)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.15)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.58)
24% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.58)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 0.99)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 562
\-------------------------
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3256; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.29)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.76)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.19)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.25)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.60)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.79)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.39)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.16)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.75)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 563
\-------------------------
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3250; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.22)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.63)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.84)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 564
\-------------------------
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3243; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.94)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.26)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.78)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.93)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.17)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.69)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.03)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.99)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.48)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.55)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.47)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 565
\-------------------------
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3237; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.88)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.19)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.14)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.51)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.50)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.54)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
77% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.71)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.06)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.20)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.35)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.38)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.29)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.70)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.10)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.25)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent drove right instead of left. (rewarded 1.25)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.40)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.42)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.42)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded -0.42)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.99)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.81)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.66)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.46)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 566
\-------------------------
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3230; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.40)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.04)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.53)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.91)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.80)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.02)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.98)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.76)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.25)
53% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 567
\-------------------------
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3224; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.04)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.58)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.97)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.72)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.51)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.54)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.20)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.38)
35% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.38)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.45)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 568
\-------------------------
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3217; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.45)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.60)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.26)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.62)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.14)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.50)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.71)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.28)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.50)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.53)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.88)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.32)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.35)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.89)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.45)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.64)
13% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 569
\-------------------------
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3211; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', 'left')
Agent followed the waypoint left. (rewarded 2.12)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.74)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.82)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.86)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 570
\-------------------------
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3205; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.87)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.67)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.98)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 0.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.25)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.94)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.54)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.13)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.19)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.48)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 571
\-------------------------
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3198; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.91)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.88)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.05)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.43)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 572
\-------------------------
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3192; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'forward')
Agent drove left instead of forward. (rewarded 1.37)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.45)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.11)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.65)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.51)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.11)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.37)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.61)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.97)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.00)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.72)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.81)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.57)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 573
\-------------------------
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3185; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.16)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.23)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent drove right instead of forward. (rewarded 1.12)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.55)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.55)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.34)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'right')
Agent followed the waypoint forward. (rewarded 1.58)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.49)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 574
\-------------------------
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3179; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.53)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 1.93)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 575
\-------------------------
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3173; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.08)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.59)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.28)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.20)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.44)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -10.98)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.27)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.89)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.34)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.20)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 576
\-------------------------
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3166; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.30)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.96)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.22)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.99)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.83)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.01)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 2.02)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 577
\-------------------------
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3160; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.81)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.69)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.03)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 578
\-------------------------
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3154; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.63)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.59)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.96)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 579
\-------------------------
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3147; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.74)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.17)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.69)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.56)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 580
\-------------------------
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3141; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.05)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.19)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.25)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.69)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.41)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.85)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
67% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.86)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.91)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.92)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 581
\-------------------------
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3135; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.01)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.42)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.51)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.36)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.07)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.87)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.64)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.72)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.94)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.26)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.40)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.57)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.90)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.92)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 582
\-------------------------
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3129; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.85)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.18)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.33)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.48)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 583
\-------------------------
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3122; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.33)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.98)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.98)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.63)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.23)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.30)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 1.79)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.90)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.76)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.60)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded 1.05)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.94)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.58)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.31)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 584
\-------------------------
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3116; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.91)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.65)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.74)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.96)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 585
\-------------------------
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3110; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 0.66)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.38)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.81)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.20)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.83)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.83)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.63)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.70)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.73)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.66)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 586
\-------------------------
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3104; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.27)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.10)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 587
\-------------------------
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3097; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.69)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'right')
Agent followed the waypoint forward. (rewarded 1.03)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 588
\-------------------------
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3091; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.57)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.22)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.70)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.30)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.64)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.39)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.75)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.81)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.60)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 589
\-------------------------
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3085; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.63)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.67)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.93)
89% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.93)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.22)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.51)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.36)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.65)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.96)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove left instead of right. (rewarded 1.68)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.24)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.23)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.62)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove forward instead of right. (rewarded 0.72)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.80)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.97)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -9.39)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
29% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.46)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.33)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded -0.02)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.35)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.28)
14% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded 0.28)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.68)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.29)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.96)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.26)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 590
\-------------------------
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3079; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.73)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.64)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 0.03)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove left instead of forward. (rewarded 0.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.78)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 0.94)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.67)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.88)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.60)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 591
\-------------------------
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3073; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.29)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.85)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.58)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.81)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.04)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.60)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.92)
67% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 592
\-------------------------
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3067; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.89)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.61)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.30)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded 0.64)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.70)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.30)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.57)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.25)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.33)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 593
\-------------------------
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3061; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.20)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.80)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.65)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.20)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 594
\-------------------------
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3054; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.77)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 595
\-------------------------
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3048; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.11)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.31)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.81)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.68)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -9.41)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.16)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.50)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 596
\-------------------------
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3042; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.92)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.41)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.57)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.60)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.56)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.41)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.34)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.04)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.02)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 597
\-------------------------
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3036; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.25)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.48)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.68)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.49)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.49)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.43)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.66)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 598
\-------------------------
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3030; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.11)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.75)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.90)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.90)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.52)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.59)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.71)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 599
\-------------------------
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3024; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.65)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.21)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.47)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.84)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.36)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.08)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.45)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 0.15)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.39)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.09)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.46)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.94)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.34)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.34)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.83)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 600
\-------------------------
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3018; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.71)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.60)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.40)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.08)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded -0.22)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.29)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.12)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.81)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.27)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded -0.62)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.55)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.15)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 601
\-------------------------
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3012; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.20)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.65)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.98)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.82)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.35)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.83)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.58)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.09)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded -0.32)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 602
\-------------------------
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3006; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.92)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.16)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.13)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.70)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.86)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.07)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.08)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 603
\-------------------------
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
Simulating trial. . .
epsilon = 0.3000; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.28)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.01)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.30)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 604
\-------------------------
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2994; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.01)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.52)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.73)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 1.53)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.27)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.08)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.84)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.36)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 1.40)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.98)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
32% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.39)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.86)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 605
\-------------------------
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2988; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent followed the waypoint right. (rewarded 2.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.52)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.07)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.36)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 606
\-------------------------
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2982; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.72)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.84)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.66)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.42)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.94)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.18)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.94)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.08)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.35)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.30)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.28)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.44)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
53% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.48)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.53)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.99)
47% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.99)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.20)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.80)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.41)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 607
\-------------------------
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2976; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.93)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.15)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.29)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.59)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.35)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.77)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.00)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.68)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.05)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.52)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.72)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.35)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.16)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.82)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 608
\-------------------------
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2970; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.04)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.86)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.67)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.26)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 1.94)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.83)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.44)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 609
\-------------------------
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2964; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.06)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.67)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.88)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 610
\-------------------------
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2958; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove forward instead of left. (rewarded 1.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.94)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.18)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.58)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.85)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.18)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -10.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.64)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.46)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.98)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.62)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.24)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.07)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.52)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.52)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.25)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.76)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.91)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.30)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 0.57)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.86)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 611
\-------------------------
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2952; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.51)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.58)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.29)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.85)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.17)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.91)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.61)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.32)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.63)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 612
\-------------------------
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2946; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.22)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.56)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent drove right instead of forward. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.78)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.71)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.57)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.44)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 0.61)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.72)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 613
\-------------------------
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2941; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.85)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.95)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.60)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.22)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.10)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.49)
28% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.49)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.93)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.32)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.29)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 614
\-------------------------
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2935; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.63)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.42)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.33)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.35)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.91)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.12)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 615
\-------------------------
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2929; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.74)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 0.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.64)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.73)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.32)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.54)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.00)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.66)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded -0.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.09)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.29)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 616
\-------------------------
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2923; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.27)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.68)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.68)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.62)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.45)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove forward instead of left. (rewarded 0.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.46)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.82)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.11)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.36)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.34)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 617
\-------------------------
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2917; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.16)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.69)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 1.88)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.60)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.40)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.17)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.72)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.14)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.68)
28% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 618
\-------------------------
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2911; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.33)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.84)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.49)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.79)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.17)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.33)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.33)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.40)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.40)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.93)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.79)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 0.94)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.75)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.31)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, 'left')
Agent drove forward instead of right. (rewarded 0.97)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.96)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.43)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.74)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.91)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.27)
23% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.27)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.13)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.84)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.78)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.97)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.81)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded -0.69)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.12)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 619
\-------------------------
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2905; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.56)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.09)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.93)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.27)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.45)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.61)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.44)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.85)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.41)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.43)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.31)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.32)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded -0.11)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.08)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.13)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.61)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.27)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.40)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.58)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.23)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
13% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.48)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.81)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.40)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 620
\-------------------------
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2900; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.30)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.66)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.05)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.69)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.63)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.22)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.38)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.46)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.99)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.62)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.72)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.28)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.93)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 621
\-------------------------
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2894; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.13)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.62)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.95)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.96)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.62)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.57)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.52)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 0.80)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.60)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.34)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.66)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent properly idled at a red light. (rewarded 2.17)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.13)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove forward instead of left. (rewarded -0.20)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.37)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.87)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.17)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.56)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 0.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 622
\-------------------------
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2888; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 1.11)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.76)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.90)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.23)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.34)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.88)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.45)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.51)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.75)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.69)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.69)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.27)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.69)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.46)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.18)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.95)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent drove left instead of right. (rewarded -0.31)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.97)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.64)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.18)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.50)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.40)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 0.08)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 623
\-------------------------
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2882; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.87)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.47)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.32)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.95)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.71)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.61)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.41)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.48)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.68)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.65)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.14)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.39)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.30)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.92)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.64)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 624
\-------------------------
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2877; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent drove right instead of forward. (rewarded 1.27)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.03)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.03)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.07)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.49)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 625
\-------------------------
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2871; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent drove left instead of right. (rewarded 1.11)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.96)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.24)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.88)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 0.85)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.36)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.80)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.21)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.29)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 0.65)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 626
\-------------------------
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2865; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.67)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.30)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.30)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove forward instead of left. (rewarded 1.77)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove right instead of left. (rewarded 0.67)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.17)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 0.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.26)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.29)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 627
\-------------------------
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2859; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.82)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.00)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.96)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.98)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 0.78)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 1.16)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.14)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.01)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.94)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'forward')
Agent drove forward instead of left. (rewarded 1.33)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.21)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.62)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded -0.20)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded -0.43)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.43)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.65)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.74)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 628
\-------------------------
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2854; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.47)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.43)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent followed the waypoint forward. (rewarded 2.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.68)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 0.09)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.87)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.55)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.09)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent drove right instead of left. (rewarded 1.54)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.39)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 1.98)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.24)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.18)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 629
\-------------------------
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2848; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.05)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.87)
68% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove right instead of left. (rewarded 1.87)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove forward instead of right. (rewarded 0.57)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.62)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.51)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.45)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.81)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.55)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.31)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.57)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.04)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.70)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.81)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.56)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 630
\-------------------------
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2842; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.83)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.10)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.13)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.80)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.52)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.53)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.07)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 631
\-------------------------
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2837; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.92)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.96)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.78)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 632
\-------------------------
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2831; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.55)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.14)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.22)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove forward instead of right. (rewarded 0.64)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.55)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.78)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.92)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.78)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded -0.39)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 0.90)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 0.89)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded -0.48)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.53)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 633
\-------------------------
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2825; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.60)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove left instead of forward. (rewarded 1.54)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.77)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.88)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.49)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.95)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 634
\-------------------------
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2820; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.16)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.93)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.54)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.59)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.62)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.42)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 635
\-------------------------
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2814; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.29)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.51)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.06)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.01)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.89)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.34)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 636
\-------------------------
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2808; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 2.45)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.42)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 0.98)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.24)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.13)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 637
\-------------------------
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2803; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.11)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.97)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.69)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.87)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.16)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.61)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.56)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.28)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.27)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.61)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded -0.30)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.42)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.83)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 638
\-------------------------
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2797; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'right')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.35)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.78)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.23)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.86)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.16)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 0.98)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.62)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.43)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 0.89)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.36)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.08)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.04)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.65)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.49)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.04)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.00)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.91)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.14)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.09)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.86)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 0.78)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 639
\-------------------------
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2792; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.49)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 640
\-------------------------
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2786; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.97)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 0.85)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 1.38)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.56)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.07)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.36)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.54)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.76)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.76)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.93)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove left instead of forward. (rewarded -0.36)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.06)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.10)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.35)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 641
\-------------------------
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2780; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.45)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.78)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.16)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.69)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.09)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.82)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.33)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 642
\-------------------------
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2775; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.39)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.92)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.49)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.22)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.19)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.32)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.26)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.14)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.84)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded 1.30)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.96)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.27)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.24)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.15)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.03)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
33% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 643
\-------------------------
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2769; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.00)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.36)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded -0.01)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.12)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.56)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 0.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.28)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 1.02)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'right', None)
Agent drove right instead of left. (rewarded 0.50)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.62)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.29)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.58)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.90)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.94)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.70)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.00)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.64)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 644
\-------------------------
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2764; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.01)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.20)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.85)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.81)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.89)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.58)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 645
\-------------------------
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2758; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.28)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.46)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.79)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.80)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.29)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.22)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.13)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.29)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 646
\-------------------------
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2753; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.07)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.09)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.01)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.95)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent drove right instead of forward. (rewarded 0.23)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'right')
Agent attempted driving left through a red light. (rewarded -10.68)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.87)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.39)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.32)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.78)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.42)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.54)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.93)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.42)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.37)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.19)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded -0.33)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.94)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.78)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.52)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 647
\-------------------------
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2747; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.11)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.96)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.59)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.99)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.72)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 648
\-------------------------
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2742; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.82)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.16)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 649
\-------------------------
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2736; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.22)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.54)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.43)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.95)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.88)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.33)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.11)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.65)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.22)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.65)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.75)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', 'right')
Agent drove right instead of forward. (rewarded -0.28)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.22)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.44)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove right instead of left. (rewarded 0.88)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.93)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 650
\-------------------------
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2731; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.04)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.81)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.67)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 0.95)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.41)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.68)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.94)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 651
\-------------------------
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2725; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.21)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.79)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.16)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.09)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.91)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.73)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'right')
Agent drove forward instead of left. (rewarded 0.10)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.88)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 0.88)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
56% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.67)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.24)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.54)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.52)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.07)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.39)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.50)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 652
\-------------------------
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2720; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.23)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent drove right instead of left. (rewarded 0.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.14)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.25)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.35)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.64)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.00)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.42)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.32)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.36)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 1.13)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.98)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.43)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.99)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.51)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 653
\-------------------------
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2714; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.41)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.48)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.40)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.41)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.73)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.54)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.56)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.87)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 654
\-------------------------
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2709; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.02)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.67)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.96)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.89)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.23)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.07)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.45)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.28)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.43)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.92)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent drove forward instead of left. (rewarded 1.45)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.75)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.45)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.74)
8% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 655
\-------------------------
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2704; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.67)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.70)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.09)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.23)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.46)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.37)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.92)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.06)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.15)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.53)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.07)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.68)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.93)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.43)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.84)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.43)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.89)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.79)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
26% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.00)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.37)
14% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 656
\-------------------------
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2698; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.11)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.39)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.94)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.52)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.77)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 657
\-------------------------
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2693; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.80)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.46)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.46)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.75)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.67)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.98)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.52)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 0.83)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.90)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 0.40)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.02)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.30)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.43)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.34)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.14)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 658
\-------------------------
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2687; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.52)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.57)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.01)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.36)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.06)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 2.10)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 659
\-------------------------
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2682; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent drove right instead of forward. (rewarded 0.99)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.51)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.81)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.84)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.86)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.67)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded -0.03)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.28)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.34)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded -0.08)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent followed the waypoint right. (rewarded 1.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 660
\-------------------------
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2677; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 2.46)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.41)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.23)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 0.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.71)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.59)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.58)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.68)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.51)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.97)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.88)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.41)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.38)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 661
\-------------------------
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2671; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.56)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.20)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.88)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 662
\-------------------------
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2666; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.83)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.82)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.60)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.14)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 663
\-------------------------
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2661; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.88)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.31)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.92)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.54)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.39)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.45)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.93)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.39)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.39)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 664
\-------------------------
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2655; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.27)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.50)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.94)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.96)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.70)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.48)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.22)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.03)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.31)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.79)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 665
\-------------------------
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2650; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.06)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.12)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.92)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.69)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.23)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.06)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.57)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 666
\-------------------------
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2645; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.02)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.50)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.57)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.77)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.86)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.76)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.37)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.82)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.74)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.09)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.72)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.69)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.86)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.45)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.07)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.81)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 0.90)
37% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 667
\-------------------------
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2639; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.02)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.88)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.55)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.57)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove right instead of left. (rewarded 1.45)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.50)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.73)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.83)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.85)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.71)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.20)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.23)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.61)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 0.34)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 668
\-------------------------
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2634; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.06)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.84)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded 1.44)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.73)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.59)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.26)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.85)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.05)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.17)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 0.99)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.30)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', None)
Agent drove forward instead of right. (rewarded 0.49)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 0.97)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.46)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 669
\-------------------------
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2629; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.08)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 2.97)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.44)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 670
\-------------------------
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2624; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 2.58)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.07)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.88)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.84)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.92)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.60)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.73)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.32)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 1.22)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', 'left')
Agent drove left instead of right. (rewarded 0.74)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.16)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.73)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.95)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove forward instead of left. (rewarded 0.47)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.19)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded -0.28)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.39)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.87)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.39)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.90)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.17)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.87)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.93)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.59)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.83)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 671
\-------------------------
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2618; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.21)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.34)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.46)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 0.80)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.93)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.16)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent followed the waypoint forward. (rewarded 2.11)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.43)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.47)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.62)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 0.82)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.82)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 0.61)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent drove forward instead of right. (rewarded 0.24)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent drove forward instead of right. (rewarded 1.18)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.17)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.07)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.33)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.29)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.02)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.58)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.12)
7% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 672
\-------------------------
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2613; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.06)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.10)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.89)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 1.40)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 0.33)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.01)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent drove right instead of forward. (rewarded 1.41)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.82)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.17)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.05)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.33)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 0.72)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.01)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.12)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.79)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.71)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.18)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.15)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 673
\-------------------------
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2608; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.43)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.02)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.06)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.25)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.81)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 674
\-------------------------
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2603; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.93)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.32)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.81)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.80)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 675
\-------------------------
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2598; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent drove forward instead of left. (rewarded 1.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.80)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.31)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 676
\-------------------------
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2592; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.93)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.08)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.19)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.72)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.99)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.70)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.13)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove right instead of forward. (rewarded 1.12)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.98)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.58)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.03)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -10.68)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.59)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.77)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.18)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.03)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 677
\-------------------------
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2587; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.44)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.26)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 678
\-------------------------
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2582; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.00)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.04)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.76)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 1.16)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.63)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.34)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.81)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.76)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.45)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.95)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.01)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded -0.13)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.37)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.57)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.58)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.76)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.10)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.11)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.38)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.17)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 0.74)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.40)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 679
\-------------------------
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2577; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.06)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.33)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.93)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.96)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.16)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'right')
Agent followed the waypoint forward. (rewarded 1.39)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.97)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.34)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.22)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 680
\-------------------------
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2572; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.61)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.59)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.64)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.51)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.37)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.62)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.24)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove right instead of left. (rewarded -0.11)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.47)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.12)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.88)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.82)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.40)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.89)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 0.70)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.00)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.82)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'right')
Agent attempted driving left through a red light. (rewarded -9.41)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 0.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 681
\-------------------------
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2567; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.63)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.10)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.36)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.72)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.52)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.42)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.97)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'forward', 'forward')
Agent drove right instead of left. (rewarded 0.65)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.94)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.70)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.76)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.18)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.60)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.49)
49% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.27)
46% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.13)
43% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.99)
40% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent drove left instead of right. (rewarded 0.20)
37% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.28)
34% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.63)
31% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.32)
29% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 0.75)
26% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded 0.45)
23% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'right')
Agent followed the waypoint left. (rewarded 1.69)
20% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
17% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.52)
14% of time remaining to reach destination.
/-------------------
| Step 30 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove forward instead of left. (rewarded 0.90)
11% of time remaining to reach destination.
/-------------------
| Step 31 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
9% of time remaining to reach destination.
/-------------------
| Step 32 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.98)
6% of time remaining to reach destination.
/-------------------
| Step 33 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.67)
3% of time remaining to reach destination.
/-------------------
| Step 34 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 682
\-------------------------
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2561; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.71)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.67)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.14)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.97)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.51)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.82)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.85)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 683
\-------------------------
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2556; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.68)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.80)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.38)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.29)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.18)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.49)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.72)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.81)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.72)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 684
\-------------------------
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2551; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.94)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.82)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.52)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.12)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.85)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
73% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.66)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.98)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.71)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.55)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 685
\-------------------------
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2546; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.12)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.82)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 1.26)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.62)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 686
\-------------------------
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2541; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.80)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.90)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.84)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.35)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 687
\-------------------------
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2536; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.83)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.22)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.57)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.39)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.56)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.13)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'right')
Agent followed the waypoint left. (rewarded 1.37)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 688
\-------------------------
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2531; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.84)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.16)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.45)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.26)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 0.34)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.03)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.49)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.14)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.82)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.26)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.98)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.72)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.72)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.20)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 689
\-------------------------
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2526; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.15)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.36)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.47)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.82)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.90)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.04)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.33)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.24)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.52)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.85)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 1.61)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.80)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.82)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.05)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.05)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
28% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.89)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.68)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.82)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.19)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
12% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.15)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.68)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.26)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.84)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 690
\-------------------------
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2521; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.58)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.79)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.75)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.58)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.31)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.02)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.44)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.69)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.17)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent drove right instead of left. (rewarded 0.97)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.14)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.50)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.09)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.84)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent drove left instead of right. (rewarded -0.86)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 691
\-------------------------
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2516; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 1.13)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent followed the waypoint forward. (rewarded 2.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.90)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.98)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.38)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.62)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.16)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'right')
Agent followed the waypoint right. (rewarded 2.40)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 692
\-------------------------
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2511; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.85)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.96)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.89)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.93)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.38)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.06)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.38)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.01)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.85)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.85)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.03)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.35)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.39)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.07)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent drove right instead of left. (rewarded 1.00)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.72)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.60)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.00)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'right')
Agent followed the waypoint right. (rewarded 1.90)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.21)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.03)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.17)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.47)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.28)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.14)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
7% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.45)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.02)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.21)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 693
\-------------------------
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2506; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.80)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.44)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.78)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.30)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.72)
65% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.72)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent drove right instead of forward. (rewarded 1.53)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 0.36)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 0.76)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.90)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'left')
Agent drove right instead of forward. (rewarded 1.07)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.80)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.77)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.25)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 694
\-------------------------
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2501; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.18)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'right')
Agent followed the waypoint right. (rewarded 1.42)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.91)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.32)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.14)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded -0.13)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.94)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 695
\-------------------------
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2496; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.72)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.53)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.30)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove forward instead of left. (rewarded 0.21)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.13)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.77)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.30)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 696
\-------------------------
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2491; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded 0.29)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.84)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.06)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.57)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.72)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.97)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.25)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.49)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.49)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.63)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.63)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.69)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
33% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 697
\-------------------------
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2486; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove left instead of forward. (rewarded 1.10)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.23)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.29)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.77)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.79)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 698
\-------------------------
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2481; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.03)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.76)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.96)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.23)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.83)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'right')
Agent followed the waypoint left. (rewarded 2.24)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.34)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.20)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.64)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 699
\-------------------------
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2476; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove right instead of forward. (rewarded 0.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.27)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.24)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.26)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 1.83)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.60)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.15)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.95)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.43)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -10.60)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.76)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent drove right instead of left. (rewarded 1.50)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent attempted driving left through a red light. (rewarded -9.85)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.08)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.89)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.21)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.73)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 700
\-------------------------
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2471; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent followed the waypoint right. (rewarded 1.17)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.06)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.87)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.62)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.82)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.82)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.94)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 701
\-------------------------
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2466; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 1.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.68)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.65)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.72)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.81)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.22)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.90)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'right', 'forward')
Agent drove left instead of forward. (rewarded 0.46)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.96)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
48% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.47)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -9.24)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.80)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove left instead of forward. (rewarded 0.20)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.57)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent attempted driving forward through a red light. (rewarded -9.94)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.55)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 702
\-------------------------
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2461; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.38)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -4.72)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.04)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent drove right instead of forward. (rewarded 0.25)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
86% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.54)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.56)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.76)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.04)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.55)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.35)
71% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 703
\-------------------------
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2456; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.36)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.66)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.91)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.59)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.78)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.62)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.69)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.73)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.33)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.28)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 704
\-------------------------
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2451; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.52)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove forward instead of right. (rewarded 0.94)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
84% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.79)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.41)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.98)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.45)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.48)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.16)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.27)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.73)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.45)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 1.51)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.40)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.52)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.18)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.60)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'forward')
Agent drove right instead of left. (rewarded 1.23)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
8% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 0.70)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.54)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.76)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 705
\-------------------------
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2446; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.92)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 1.36)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.05)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.59)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.25)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.75)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.80)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.04)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.95)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.28)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.63)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.70)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.05)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove right instead of forward. (rewarded 1.29)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.24)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.02)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
33% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 706
\-------------------------
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2441; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.34)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.98)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.78)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove forward instead of right. (rewarded 1.12)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.76)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.83)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.30)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -10.28)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.50)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.34)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.92)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.68)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.22)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 1.10)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.04)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.89)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.25)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.72)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.22)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.83)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.86)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 707
\-------------------------
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2437; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 1.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.42)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.06)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.06)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.37)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.89)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 708
\-------------------------
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2432; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.08)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent drove right instead of forward. (rewarded 1.22)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.46)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.57)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.66)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent drove left instead of forward. (rewarded 1.62)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.66)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 2.28)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 2.51)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 709
\-------------------------
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2427; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.88)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.58)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.75)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.66)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.44)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.53)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.09)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 0.90)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.85)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.54)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.61)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent drove right instead of left. (rewarded 1.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.53)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.44)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.91)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.94)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 1.57)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.55)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 710
\-------------------------
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2422; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.69)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.68)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.93)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 1.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.36)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.15)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.07)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.12)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.90)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.44)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.36)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.43)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 0.50)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.67)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.28)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.74)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 711
\-------------------------
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2417; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.35)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
93% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.06)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.00)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.15)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.63)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 712
\-------------------------
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2412; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.46)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.52)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.33)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.23)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 713
\-------------------------
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2407; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.64)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.57)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent attempted driving left through a red light. (rewarded -9.97)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent attempted driving forward through a red light. (rewarded -9.15)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent followed the waypoint forward. (rewarded 2.80)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.48)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.10)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.86)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.32)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.22)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent attempted driving left through a red light. (rewarded -9.22)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.16)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.52)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.61)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 1.14)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.32)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.23)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.01)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.45)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.43)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
17% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.44)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
13% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.05)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 714
\-------------------------
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2403; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.08)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -9.39)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.27)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.02)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.26)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.36)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.92)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.10)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.93)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 1.21)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.71)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 1.25)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 0.55)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.37)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 1.16)
16% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 715
\-------------------------
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2398; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.98)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 2.42)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.48)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.88)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.76)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.68)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.99)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.86)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.74)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.33)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.92)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.40)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.78)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.59)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.84)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 716
\-------------------------
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2393; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', 'left')
Agent followed the waypoint right. (rewarded 1.04)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 1.82)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.67)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.38)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.31)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.79)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.94)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 717
\-------------------------
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2388; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 0.18)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.62)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.24)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.80)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.44)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.44)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.46)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.64)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.16)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.54)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.52)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.13)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.20)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.57)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.03)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.88)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.81)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 718
\-------------------------
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2384; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.06)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.68)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.51)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.60)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.74)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.27)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.92)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.84)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.06)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.44)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.66)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.08)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.86)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.75)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 719
\-------------------------
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2379; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', 'left')
Agent drove forward instead of left. (rewarded 0.70)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.27)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.58)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.17)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.07)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.38)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.59)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.38)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.60)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 720
\-------------------------
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2374; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.85)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.64)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.03)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.14)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
50% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 721
\-------------------------
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2369; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.38)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.27)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.89)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.19)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.68)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.70)
65% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 2.70)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.51)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.15)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.89)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.42)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.83)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.88)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'forward')
Agent drove left instead of right. (rewarded 1.48)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.49)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.05)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.73)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 722
\-------------------------
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2365; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.32)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.81)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.60)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.99)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.67)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.82)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.36)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 723
\-------------------------
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2360; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.10)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.11)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.76)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.46)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.28)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.60)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.21)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.01)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.17)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.43)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.02)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.73)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.97)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.35)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 0.81)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.36)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.45)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.26)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded -0.18)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.27)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 0.26)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.24)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 724
\-------------------------
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2355; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.50)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.59)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.10)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.58)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.39)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.50)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent followed the waypoint forward. (rewarded 1.96)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.55)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
30% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -9.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.62)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.64)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 725
\-------------------------
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2350; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'left')
Agent properly idled at a red light. (rewarded 1.37)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.78)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 1.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.46)
75% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 726
\-------------------------
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2346; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.33)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent attempted driving forward through a red light. (rewarded -10.45)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.25)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.69)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.60)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.25)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.43)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.86)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.33)
40% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.33)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.48)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 727
\-------------------------
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2341; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.01)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.62)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.62)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.10)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.93)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.82)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.00)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.34)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.23)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.24)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 728
\-------------------------
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2336; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 0.81)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.01)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.47)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.27)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.61)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent drove left instead of right. (rewarded 1.82)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent drove right instead of forward. (rewarded 0.30)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.11)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.11)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.94)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', 'right')
Agent followed the waypoint left. (rewarded 2.70)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.57)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.89)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.09)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.03)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.92)
37% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 729
\-------------------------
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2332; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.24)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 2.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded 1.50)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.02)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.30)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.07)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.36)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.10)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 730
\-------------------------
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2327; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.98)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.35)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.08)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.38)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.85)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.50)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.18)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.83)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.81)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.12)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.90)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.44)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.27)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.65)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.59)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.27)
28% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 731
\-------------------------
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2322; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.23)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 0.20)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 1.29)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.04)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.01)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.47)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.97)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.73)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 1.16)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
30% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.12)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
20% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.52)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.82)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.66)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 732
\-------------------------
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2318; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.76)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.50)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.00)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.79)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.40)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.93)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.26)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.55)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.06)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.75)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.83)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.03)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 0.70)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.56)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.33)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.77)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.06)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove right instead of left. (rewarded 0.13)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 2.26)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.01)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.36)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 733
\-------------------------
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2313; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 1.96)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 2.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.83)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.75)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.56)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.95)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.05)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.64)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.67)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.71)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.11)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 734
\-------------------------
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2308; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.98)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.66)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent followed the waypoint forward. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.78)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.36)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -9.16)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.39)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.84)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.10)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded -0.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.82)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.10)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.85)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.93)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.95)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 735
\-------------------------
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2304; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.09)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.09)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.92)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 2.37)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.65)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'right', None)
Agent properly idled at a red light. (rewarded 1.80)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.13)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.79)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.60)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.94)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded -0.03)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'forward', None)
Agent drove forward instead of left. (rewarded 1.26)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.97)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -5.28)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.29)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.11)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.21)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.60)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.12)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.46)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 736
\-------------------------
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2299; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.24)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.25)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.97)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.42)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.12)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.04)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.38)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 737
\-------------------------
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2295; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.05)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 2.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.21)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.18)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.67)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.19)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 738
\-------------------------
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2290; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.12)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -10.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.75)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.00)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.78)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 1.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.15)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 739
\-------------------------
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2286; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.66)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.79)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.85)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.47)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.03)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.33)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.39)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 2.37)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.94)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.62)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.47)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.28)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.67)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.74)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 740
\-------------------------
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2281; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.47)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.51)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.29)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.16)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.14)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.79)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.89)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.62)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.38)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent idled at a green light with no oncoming traffic. (rewarded -5.20)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.65)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'right')
Agent followed the waypoint forward. (rewarded 1.48)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.80)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 741
\-------------------------
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2276; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 2.34)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'right')
Agent drove left instead of right. (rewarded 0.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.65)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.16)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.48)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.17)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.83)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.88)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.41)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.38)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.74)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.04)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'left')
Agent attempted driving right through traffic and cause a minor accident. (rewarded -19.53)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.33)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 742
\-------------------------
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2272; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 0.76)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.76)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.78)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.13)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.16)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.44)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.94)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.53)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.76)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.22)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 743
\-------------------------
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2267; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.99)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.30)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.42)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.91)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.64)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.17)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.42)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.94)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.46)
67% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', 'left')
Agent drove right instead of forward. (rewarded 1.46)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.51)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.37)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded -0.12)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'forward', None)
Agent drove right instead of left. (rewarded 0.89)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.02)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.58)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded -0.27)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, 'forward')
Agent drove forward instead of right. (rewarded -0.27)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 1.60)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.45)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.42)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.89)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.24)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.18)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.18)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.27)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.58)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 0.56)
7% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'right', None, 'right')
Agent followed the waypoint right. (rewarded 0.56)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.42)
3% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 744
\-------------------------
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2263; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.03)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.29)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.74)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 1.71)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.28)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent drove forward instead of left. (rewarded 0.20)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.75)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.48)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.22)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.45)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
52% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.23)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.17)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.35)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded -0.16)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.07)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.95)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.64)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded 0.41)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 0.97)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 1.70)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.67)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'right', None)
Agent followed the waypoint left. (rewarded 1.39)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.60)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 745
\-------------------------
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2258; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.98)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.46)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.83)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', 'left')
Agent followed the waypoint forward. (rewarded 1.50)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.12)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.31)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.06)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded -0.06)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.87)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving forward through a red light. (rewarded -10.90)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.00)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 0.29)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.28)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 746
\-------------------------
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2254; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.63)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.25)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.81)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.18)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.35)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.47)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 0.05)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded -0.08)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.00)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.18)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.04)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.26)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.06)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 747
\-------------------------
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2249; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'forward')
Agent drove right instead of left. (rewarded 0.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.37)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.83)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.19)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'left')
Agent drove left instead of forward. (rewarded 1.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.63)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 2.08)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.51)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 748
\-------------------------
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2245; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.17)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.06)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.57)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.23)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.50)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.86)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.57)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'forward', 'left')
Agent drove right instead of left. (rewarded 0.93)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.86)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.24)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.15)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 0.89)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.59)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -9.59)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.55)
36% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 749
\-------------------------
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2240; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.84)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.30)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.09)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.99)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.22)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.77)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 0.99)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.28)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.82)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.11)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.46)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.38)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 750
\-------------------------
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2236; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.50)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.55)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.63)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 751
\-------------------------
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2231; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.31)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.03)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.47)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.98)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.69)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.72)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.21)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.90)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.83)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.25)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.73)
44% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 752
\-------------------------
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2227; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', 'forward')
Agent drove left instead of right. (rewarded 0.09)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.84)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.27)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.51)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.40)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.73)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.76)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent followed the waypoint forward. (rewarded 1.50)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.93)
20% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 753
\-------------------------
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2222; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.75)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.48)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.35)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.36)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove right instead of left. (rewarded 1.34)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.46)
60% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.46)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.91)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', 'forward')
Agent drove right instead of forward. (rewarded 0.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 1.16)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.52)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 0.46)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.09)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.03)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 754
\-------------------------
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2218; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.77)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.81)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.28)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.11)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.08)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 2.80)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 755
\-------------------------
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2214; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent drove right instead of forward. (rewarded 0.75)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 2.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.13)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.72)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.89)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded -0.17)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', None)
Agent followed the waypoint right. (rewarded 1.79)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.60)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 756
\-------------------------
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2209; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.28)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.57)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.24)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.24)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.89)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent drove right instead of forward. (rewarded 0.73)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.15)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.37)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.60)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.20)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.31)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 757
\-------------------------
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2205; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.81)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.71)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.56)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.78)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.80)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.85)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.15)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.59)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.19)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.07)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.78)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 0.61)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.09)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 758
\-------------------------
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2200; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 1.35)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.72)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving left through a red light. (rewarded -10.31)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.82)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.50)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.95)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded -0.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 759
\-------------------------
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2196; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.16)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.85)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.21)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.21)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.87)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.31)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.53)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.89)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.79)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.07)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.72)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -9.97)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'right')
Agent drove forward instead of right. (rewarded 0.74)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.77)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.49)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.15)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.34)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
32% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 760
\-------------------------
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2191; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.73)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.39)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.32)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.54)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.43)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.40)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.28)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.78)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.27)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 1.37)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 1.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded -0.43)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.70)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.11)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.08)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.95)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded -0.56)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 761
\-------------------------
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2187; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 0.10)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent drove forward instead of left. (rewarded 0.65)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.35)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.77)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'left')
Agent drove forward instead of left. (rewarded 1.77)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 1.00)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.85)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.51)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'right', None)
Agent attempted driving forward through a red light. (rewarded -10.44)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.25)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.03)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.45)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.26)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.92)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.81)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove left instead of forward. (rewarded 1.21)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.68)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.94)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.71)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.01)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.50)
17% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 762
\-------------------------
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2183; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent drove left instead of right. (rewarded 0.00)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.66)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.23)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.58)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.60)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.81)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.76)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.86)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 1.71)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.45)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.05)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 763
\-------------------------
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2178; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.24)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.43)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.89)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 764
\-------------------------
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2174; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.28)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.42)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.76)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.02)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.18)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.73)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.52)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.68)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.11)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.79)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent drove left instead of right. (rewarded 1.21)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.07)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent attempted driving left through a red light. (rewarded -10.70)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.28)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.71)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 1.06)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.82)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.25)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.13)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.92)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 0.93)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent followed the waypoint right. (rewarded 0.50)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', None)
Agent properly idled at a red light. (rewarded 0.29)
4% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 0.64)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 765
\-------------------------
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2170; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.53)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.12)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.95)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.52)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.79)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 766
\-------------------------
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2165; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', 'left')
Agent drove forward instead of left. (rewarded 1.20)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.86)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.51)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.08)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.98)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.40)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.50)
72% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.50)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.57)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.66)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.37)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.74)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.52)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove right instead of forward. (rewarded -0.14)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.75)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.44)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 0.62)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.66)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.75)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.73)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 1.95)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 0.71)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.80)
12% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 767
\-------------------------
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2161; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.07)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving forward through a red light. (rewarded -10.64)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.62)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.35)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.85)
65% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent attempted driving left through a red light. (rewarded -10.85)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.08)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', None)
Agent attempted driving left through a red light. (rewarded -9.08)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 0.81)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 2.11)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.15)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.70)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 768
\-------------------------
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2157; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.17)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.15)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.48)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.00)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.85)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.33)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.82)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.91)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'right', None)
Agent followed the waypoint forward. (rewarded 1.02)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.59)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.88)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 769
\-------------------------
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2152; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.12)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 2.30)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.53)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.70)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.15)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.58)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.59)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 2.38)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.14)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.19)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.91)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 770
\-------------------------
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2148; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.04)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.79)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 2.61)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.39)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.01)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.91)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.57)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.71)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.85)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.65)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.15)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 771
\-------------------------
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2144; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', None)
Agent followed the waypoint right. (rewarded 2.64)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.22)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.41)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.57)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.15)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.60)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent attempted driving forward through a red light. (rewarded -9.98)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.68)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.07)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.37)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.34)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.80)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.91)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.65)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.58)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.29)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 772
\-------------------------
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2140; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'right')
Agent followed the waypoint right. (rewarded 1.57)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.53)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'left', None)
Agent drove forward instead of right. (rewarded 1.51)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent drove left instead of right. (rewarded 0.97)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 0.84)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.44)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.31)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.87)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.53)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.11)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'right', None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.79)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 1.12)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.47)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, 'forward')
Agent properly idled at a red light. (rewarded 2.34)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.36)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.24)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.84)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 773
\-------------------------
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2135; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.65)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.81)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.46)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.43)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.63)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 0.83)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.17)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.53)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.26)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.54)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.41)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.35)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.76)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.59)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 0.60)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.01)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 1.09)
15% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 774
\-------------------------
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2131; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent drove right instead of forward. (rewarded 0.93)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.45)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.92)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.50)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.95)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.27)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.44)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.41)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 775
\-------------------------
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2127; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.69)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.36)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 776
\-------------------------
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2122; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.49)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.07)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.33)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'left')
Agent followed the waypoint forward. (rewarded 1.10)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.38)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.86)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 1.52)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 0.29)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.44)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 1.71)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.05)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.44)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.69)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.09)
47% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 777
\-------------------------
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2118; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.44)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.59)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.39)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.43)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.78)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.89)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.66)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.74)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.34)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.11)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.45)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.26)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.70)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.86)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.26)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.19)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.47)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.16)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'green', 'right', 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.79)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 778
\-------------------------
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2114; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.92)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.61)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 2.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.79)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.77)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.08)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.30)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.34)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.04)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.23)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 0.65)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 779
\-------------------------
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2110; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.28)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.97)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.33)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.03)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.05)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.83)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving forward through a red light. (rewarded -9.50)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.99)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 0.99)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.72)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.52)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 0.96)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 780
\-------------------------
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2106; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.57)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 0.99)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.64)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'left')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 1.27)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.44)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.89)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.46)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.30)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.67)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 0.90)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.66)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.65)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 781
\-------------------------
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2101; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 1.51)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.38)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.21)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.28)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.56)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.41)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.15)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 1.79)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 782
\-------------------------
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2097; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.31)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.40)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'forward')
Agent drove right instead of left. (rewarded 1.33)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.50)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.23)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.20)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.21)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.24)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.06)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.63)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 2.44)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.42)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.39)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.13)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.04)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.27)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.93)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.99)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 783
\-------------------------
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2093; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 2.65)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.86)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.92)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent drove forward instead of left. (rewarded -0.02)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'forward')
Agent drove right instead of left. (rewarded 0.82)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.17)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent followed the waypoint right. (rewarded 1.67)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.63)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 1.14)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.48)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, 'forward')
Agent followed the waypoint right. (rewarded 1.24)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.83)
48% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 784
\-------------------------
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2089; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.55)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', 'right')
Agent properly idled at a red light. (rewarded 1.29)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', None)
Agent properly idled at a red light. (rewarded 2.29)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.89)
87% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 0.89)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 1.99)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.76)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.80)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.00)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.16)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 1.70)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.67)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.76)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.83)
57% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 785
\-------------------------
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2085; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', 'left')
Agent followed the waypoint right. (rewarded 1.99)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.25)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', 'left')
Agent drove left instead of forward. (rewarded 0.19)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.29)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 1.16)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.18)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.19)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.13)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.54)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.96)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.54)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.61)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.95)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
44% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.96)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'forward')
Agent drove forward instead of right. (rewarded -0.20)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 0.68)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent drove left instead of right. (rewarded 0.81)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.14)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 0.54)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.06)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.04)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent followed the waypoint forward. (rewarded 1.41)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded -0.59)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.85)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 786
\-------------------------
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2080; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 1.43)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.73)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.72)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.97)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.99)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded 1.19)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.05)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.81)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.11)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded -0.21)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.27)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.53)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded -0.49)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.25)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', None)
Agent followed the waypoint right. (rewarded 1.35)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.69)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'left')
Agent followed the waypoint forward. (rewarded 1.38)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 787
\-------------------------
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2076; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'right')
Agent idled at a green light with no oncoming traffic. (rewarded -4.39)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.70)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 1.96)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.23)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 2.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.76)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.20)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.05)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 788
\-------------------------
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2072; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'left')
Agent properly idled at a red light. (rewarded 1.47)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 1.18)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.35)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.01)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.40)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.68)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.07)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.77)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent followed the waypoint forward. (rewarded 2.40)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.89)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.64)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.00)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.70)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.69)
43% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 789
\-------------------------
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2068; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 1.41)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.51)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.03)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.59)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.95)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.36)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.63)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving left through a red light. (rewarded -9.88)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.64)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 1.97)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.26)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'right')
Agent drove right instead of forward. (rewarded 1.00)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.40)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.16)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 790
\-------------------------
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2064; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.22)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.22)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.47)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.05)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.04)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.49)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.65)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.33)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.20)
55% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 791
\-------------------------
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2060; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.90)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.27)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', None, 'right')
Agent drove forward instead of left. (rewarded 0.89)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 1.15)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.32)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 792
\-------------------------
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2056; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent properly idled at a red light. (rewarded 1.99)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'forward')
Agent properly idled at a red light. (rewarded 2.08)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.30)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.87)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.32)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.41)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 1.23)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded -0.11)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'forward', 'forward', 'forward')
Agent drove forward instead of right. (rewarded -0.11)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.99)
52% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 793
\-------------------------
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2052; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.88)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.18)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'left')
Agent followed the waypoint right. (rewarded 1.30)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.46)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.52)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.10)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.11)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.66)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent drove right instead of left. (rewarded 0.53)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 0.75)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.18)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.56)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.75)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.29)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent attempted driving forward through a red light. (rewarded -9.18)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'forward')
Agent followed the waypoint left. (rewarded 1.27)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 794
\-------------------------
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2047; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.86)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 2.48)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.39)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.19)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.20)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.14)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 2.33)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 1.15)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
64% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 795
\-------------------------
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2043; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.01)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', 'left')
Agent properly idled at a red light. (rewarded 2.59)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.02)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.63)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.44)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.93)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 0.94)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 0.90)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.62)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 0.90)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', 'forward', None)
Agent drove right instead of forward. (rewarded 1.55)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.52)
30% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent drove forward instead of left. (rewarded 0.52)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.56)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'forward', None)
Agent drove forward instead of left. (rewarded -0.12)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.21)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.18)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.06)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent drove right instead of forward. (rewarded -0.23)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 796
\-------------------------
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2039; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent properly idled at a red light. (rewarded 1.52)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.47)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'left')
Agent drove right instead of left. (rewarded 1.49)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.76)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent followed the waypoint right. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.31)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.39)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'left')
Agent followed the waypoint forward. (rewarded 2.73)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', 'forward')
Agent followed the waypoint forward. (rewarded 2.62)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 2.57)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 797
\-------------------------
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2035; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.91)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.06)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -5.95)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.15)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.40)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.94)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.23)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.49)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.25)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.49)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 798
\-------------------------
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2031; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.54)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', None)
Agent drove right instead of forward. (rewarded 0.39)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.20)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 799
\-------------------------
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2027; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 2.26)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.26)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.55)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.38)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.66)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.94)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.71)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.08)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'forward')
Agent followed the waypoint left. (rewarded 0.91)
64% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 800
\-------------------------
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2023; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.77)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'left', None)
Agent drove right instead of left. (rewarded 1.23)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.93)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', 'forward')
Agent properly idled at a red light. (rewarded 1.01)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'forward')
Agent drove forward instead of left. (rewarded -0.07)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', 'left', None)
Agent drove right instead of left. (rewarded 0.78)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', None, 'right', 'left')
Agent drove forward instead of right. (rewarded 1.47)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.72)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.63)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, 'left')
Agent drove forward instead of right. (rewarded -0.24)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 0.94)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'left')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.44)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.03)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.76)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded -0.55)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.19)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.10)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 801
\-------------------------
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2019; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'right', 'forward')
Agent followed the waypoint right. (rewarded 2.55)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.11)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.54)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove left instead of forward. (rewarded 0.27)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'forward')
Agent followed the waypoint right. (rewarded 2.55)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.82)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.95)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'right', 'left', 'forward')
Agent properly idled at a red light. (rewarded 0.95)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.64)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.71)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.76)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.16)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.05)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 1.68)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.06)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.43)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.58)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.74)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.38)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', None)
Agent followed the waypoint forward. (rewarded 2.00)
33% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 802
\-------------------------
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2015; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent attempted driving left through a red light. (rewarded -9.02)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', None, None)
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.43)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.59)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.45)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.05)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.52)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.43)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent drove left instead of forward. (rewarded 1.44)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 1.44)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 803
\-------------------------
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2011; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent properly idled at a red light. (rewarded 2.48)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.03)
94% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.25)
91% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'left')
Agent idled at a green light with no oncoming traffic. (rewarded -4.07)
89% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.62)
86% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.30)
83% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'left')
Agent properly idled at a red light. (rewarded 1.00)
80% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'left', None)
Agent followed the waypoint left. (rewarded 1.48)
77% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.62)
74% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'forward')
Agent properly idled at a red light. (rewarded 1.95)
71% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.77)
69% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent drove right instead of forward. (rewarded 1.17)
66% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.72)
63% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 1.09)
60% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.12)
57% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.71)
54% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.45)
51% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.11)
49% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 804
\-------------------------
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2007; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', None)
Agent properly idled at a red light. (rewarded 2.39)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.61)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', 'right')
Agent properly idled at a red light. (rewarded 2.78)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', None)
Agent drove right instead of left. (rewarded 1.55)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, 'forward')
Agent followed the waypoint right. (rewarded 2.21)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.24)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.41)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.29)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 1.72)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.13)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.25)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded -0.02)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.69)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'forward')
Agent properly idled at a red light. (rewarded 1.82)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent attempted driving left through a red light. (rewarded -10.18)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.51)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.10)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 1.89)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.55)
37% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 805
\-------------------------
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
Simulating trial. . .
epsilon = 0.2003; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.83)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.92)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.27)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent drove right instead of left. (rewarded 1.16)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, 'right')
Agent drove right instead of left. (rewarded 1.50)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.29)
65% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.29)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove right instead of forward. (rewarded 1.55)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.08)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, 'left')
Agent followed the waypoint left. (rewarded 2.48)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.27)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 806
\-------------------------
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1999; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'left')
Agent followed the waypoint right. (rewarded 2.77)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.15)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', 'forward')
Agent followed the waypoint right. (rewarded 2.46)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'forward')
Agent drove right instead of forward. (rewarded 1.02)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'left')
Agent followed the waypoint left. (rewarded 2.58)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.10)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.55)
72% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 807
\-------------------------
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1995; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 2.62)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.51)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.04)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.73)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded -0.08)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'forward', None)
Agent followed the waypoint right. (rewarded 1.61)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.66)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -39.41)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.04)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.50)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 808
\-------------------------
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1991; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.02)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.53)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 1.93)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.43)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.42)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
76% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.74)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.67)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'forward', 'right')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.92)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent drove left instead of forward. (rewarded -0.03)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.73)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.73)
56% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 809
\-------------------------
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1987; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.94)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.34)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.38)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.34)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.09)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.36)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.43)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.44)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.66)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 0.86)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.20)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent drove forward instead of left. (rewarded 0.87)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded 0.49)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.51)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', 'forward')
Agent followed the waypoint right. (rewarded 0.65)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 0.58)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 0.63)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 810
\-------------------------
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1983; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'right', 'forward')
Agent properly idled at a red light. (rewarded 2.31)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.21)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.95)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent attempted driving left through a red light. (rewarded -9.89)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent drove right instead of forward. (rewarded 1.37)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'right', None, None)
Agent properly idled at a red light. (rewarded 1.17)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.45)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, 'left')
Agent drove forward instead of left. (rewarded 1.24)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.39)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.89)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.73)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'forward', 'right', None)
Agent drove forward instead of left. (rewarded 0.75)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
57% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.58)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.89)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
50% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.97)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 0.86)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.10)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 0.59)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.10)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.30)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.28)
30% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent drove right instead of forward. (rewarded 0.76)
27% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent drove right instead of left. (rewarded 0.51)
23% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.63)
20% of time remaining to reach destination.
/-------------------
| Step 24 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.59)
17% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', 'right', None)
Agent properly idled at a red light. (rewarded 0.59)
17% of time remaining to reach destination.
/-------------------
| Step 25 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent drove right instead of left. (rewarded -0.64)
13% of time remaining to reach destination.
/-------------------
| Step 26 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 0.36)
10% of time remaining to reach destination.
/-------------------
| Step 27 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'right')
Agent attempted driving left through a red light. (rewarded -10.75)
7% of time remaining to reach destination.
/-------------------
| Step 28 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.14)
3% of time remaining to reach destination.
/-------------------
| Step 29 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 0.70)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 811
\-------------------------
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1979; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'left', 'left', 'left')
Agent followed the waypoint right. (rewarded 2.82)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.69)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving left through a red light. (rewarded -10.85)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.24)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent attempted driving forward through a red light. (rewarded -10.37)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.24)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', None)
Agent followed the waypoint forward. (rewarded 2.56)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.13)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 812
\-------------------------
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1975; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 2.51)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('right', 'red', 'right', 'right', None)
Agent followed the waypoint right. (rewarded 2.62)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 1.75)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.28)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.02)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.75)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', 'right')
Agent drove forward instead of left. (rewarded 1.70)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.64)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -9.35)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.76)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.14)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.94)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.74)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 2.51)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.52)
25% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 813
\-------------------------
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1971; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.36)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', None, 'right')
Agent properly idled at a red light. (rewarded 1.85)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.29)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 2.55)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.47)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'right', None)
Agent drove right instead of forward. (rewarded 0.76)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.15)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.13)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -9.72)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'right')
Agent followed the waypoint left. (rewarded 2.74)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.47)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.32)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'right', 'right')
Agent properly idled at a red light. (rewarded 2.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.12)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent attempted driving left through traffic and cause a minor accident. (rewarded -19.39)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.38)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
15% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.53)
10% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.39)
5% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent drove forward instead of right. (rewarded 0.43)
0% of time remaining to reach destination.
Trial Aborted!
Agent did not reach the destination.
/-------------------------
| Training trial 814
\-------------------------
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1967; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', 'left')
Agent followed the waypoint left. (rewarded 2.11)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 1.18)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.89)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.72)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.64)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.22)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.09)
68% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 815
\-------------------------
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1963; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.13)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', None)
Agent followed the waypoint left. (rewarded 2.21)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.17)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.88)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.55)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.27)
70% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 816
\-------------------------
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1959; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', 'forward')
Agent followed the waypoint left. (rewarded 1.59)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.98)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'forward')
Agent attempted driving forward through a red light with traffic and cause a major accident. (rewarded -40.48)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.39)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.68)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.96)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 0.99)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.72)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.86)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'right')
Agent followed the waypoint forward. (rewarded 2.23)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', None)
Agent drove right instead of forward. (rewarded 1.70)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'left')
Agent properly idled at a red light. (rewarded 2.02)
60% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'forward', None)
Agent attempted driving right through traffic and cause a minor accident. (rewarded -20.58)
57% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.08)
53% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
50% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', None)
Agent properly idled at a red light. (rewarded 1.99)
47% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', 'forward', None)
Agent drove right instead of left. (rewarded 0.70)
43% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.46)
40% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.44)
37% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', 'left', None)
Agent idled at a green light with no oncoming traffic. (rewarded -4.75)
33% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.46)
30% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 817
\-------------------------
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1955; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 1.58)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'forward', 'right', 'left')
Agent drove right instead of forward. (rewarded 0.64)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.36)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
84% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.27)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.70)
80% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 818
\-------------------------
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1951; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', 'left', 'right')
Agent attempted driving forward through a red light. (rewarded -10.88)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.01)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.77)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'forward', None, None)
Agent idled at a green light with no oncoming traffic. (rewarded -5.42)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.70)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.77)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.65)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.38)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.61)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.08)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.06)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.05)
52% of time remaining to reach destination.
Agent previous state: ('right', 'green', 'left', None, 'left')
Agent drove forward instead of right. (rewarded 0.05)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 1.26)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.43)
44% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'forward', 'right', None)
Agent followed the waypoint right. (rewarded 2.43)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.35)
40% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 819
\-------------------------
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1948; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, 'forward')
Agent followed the waypoint forward. (rewarded 2.41)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.05)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.80)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.75)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.39)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', None)
Agent followed the waypoint right. (rewarded 1.82)
65% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 820
\-------------------------
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1944; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.20)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.41)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent attempted driving forward through a red light. (rewarded -10.97)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.71)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'green', None, 'left', None)
Agent drove forward instead of left. (rewarded 0.86)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 1.65)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.63)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.18)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.66)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent drove left instead of forward. (rewarded 0.46)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent attempted driving left through a red light. (rewarded -10.52)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 2.19)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.54)
35% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 821
\-------------------------
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1940; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, 'left', 'right')
Agent followed the waypoint right. (rewarded 2.03)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.91)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent followed the waypoint forward. (rewarded 2.42)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 1.19)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.30)
76% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 822
\-------------------------
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1936; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 2.14)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.23)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 1.38)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', 'left', None)
Agent properly idled at a red light. (rewarded 2.54)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 2.33)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
76% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.52)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent attempted driving left through traffic and cause a minor accident. (rewarded -20.14)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 2.58)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.40)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'left')
Agent drove left instead of forward. (rewarded 0.29)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', None)
Agent followed the waypoint right. (rewarded 1.21)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove right instead of forward. (rewarded 0.26)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.04)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.84)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, None)
Agent followed the waypoint right. (rewarded 2.36)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, 'left')
Agent followed the waypoint right. (rewarded 1.21)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.59)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', 'forward', 'forward')
Agent properly idled at a red light. (rewarded 2.43)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.13)
24% of time remaining to reach destination.
/-------------------
| Step 19 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.80)
20% of time remaining to reach destination.
/-------------------
| Step 20 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.47)
16% of time remaining to reach destination.
/-------------------
| Step 21 Results
\-------------------
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
Agent previous state: ('right', 'red', None, None, None)
Agent followed the waypoint right. (rewarded 1.82)
12% of time remaining to reach destination.
/-------------------
| Step 22 Results
\-------------------
Agent previous state: ('left', 'red', 'left', 'right', None)
Agent properly idled at a red light. (rewarded 1.06)
8% of time remaining to reach destination.
/-------------------
| Step 23 Results
\-------------------
Agent previous state: ('left', 'green', 'left', None, None)
Agent followed the waypoint left. (rewarded 0.45)
4% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 823
\-------------------------
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1932; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'forward')
Agent followed the waypoint right. (rewarded 2.79)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, 'right')
Agent drove right instead of forward. (rewarded 1.31)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent drove forward instead of left. (rewarded 1.18)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'forward')
Agent followed the waypoint left. (rewarded 1.05)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.94)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.22)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.76)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 2.49)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'left', None)
Agent properly idled at a red light. (rewarded 2.47)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 2.42)
67% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 824
\-------------------------
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1928; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'left', None)
Agent drove forward instead of right. (rewarded 0.37)
96% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.51)
92% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.17)
88% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.68)
84% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.78)
80% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 1.55)
76% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', None)
Agent followed the waypoint forward. (rewarded 2.58)
72% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent drove right instead of forward. (rewarded 0.57)
68% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', 'left')
Agent drove forward instead of left. (rewarded 1.04)
64% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'left')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -40.18)
60% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'right', 'left', 'forward')
Agent drove right instead of left. (rewarded 0.43)
56% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('right', 'green', None, None, 'forward')
Agent followed the waypoint right. (rewarded 0.95)
52% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'left')
Agent properly idled at a red light. (rewarded 1.62)
48% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', None, None)
Agent drove forward instead of right. (rewarded 1.34)
44% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'left')
Agent followed the waypoint right. (rewarded 2.14)
40% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', 'right', 'right')
Agent drove right instead of forward. (rewarded -0.16)
36% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.41)
32% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', 'forward', None)
Agent followed the waypoint forward. (rewarded 2.42)
28% of time remaining to reach destination.
/-------------------
| Step 18 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.87)
24% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 825
\-------------------------
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1924; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, 'forward')
Agent drove right instead of forward. (rewarded 1.11)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 0.26)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.85)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, None, None)
Agent drove right instead of left. (rewarded 1.85)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.00)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, 'forward')
Agent followed the waypoint forward. (rewarded 2.37)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.82)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', 'forward', None, 'right')
Agent drove right instead of left. (rewarded 1.46)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'right')
Agent drove left instead of forward. (rewarded 0.49)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'right', 'left')
Agent followed the waypoint right. (rewarded 2.60)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'forward', 'forward', 'left')
Agent drove forward instead of right. (rewarded 0.33)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, 'forward', None)
Agent followed the waypoint right. (rewarded 1.11)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 826
\-------------------------
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1920; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'green', None, 'forward', 'left')
Agent followed the waypoint left. (rewarded 1.87)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'left', None)
Agent followed the waypoint forward. (rewarded 2.08)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
Agent previous state: ('left', 'red', 'left', None, None)
Agent drove right instead of left. (rewarded 1.07)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', None, None)
Agent followed the waypoint forward. (rewarded 1.78)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'left')
Agent properly idled at a red light. (rewarded 1.74)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.02)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 1.26)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'forward', 'left')
Agent drove left instead of forward. (rewarded 1.54)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 1.79)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'green', 'left', None, None)
Agent followed the waypoint right. (rewarded 1.98)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', 'left', 'left', None)
Agent followed the waypoint left. (rewarded 2.47)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 827
\-------------------------
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1917; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'forward', 'left', 'forward')
Agent followed the waypoint forward. (rewarded 1.69)
97% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.81)
93% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.21)
90% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, 'right')
Agent properly idled at a red light. (rewarded 2.09)
87% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, None)
Agent properly idled at a red light. (rewarded 1.19)
83% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'right')
Agent drove right instead of forward. (rewarded 1.80)
80% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.96)
77% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 1.63)
73% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.17)
70% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 0.94)
67% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
Agent previous state: ('right', 'green', None, None, None)
Agent followed the waypoint right. (rewarded 0.91)
63% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, None)
Agent followed the waypoint forward. (rewarded 1.11)
60% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 828
\-------------------------
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1913; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', 'left', 'left')
Agent followed the waypoint forward. (rewarded 2.97)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('forward', 'green', 'right', None, None)
Agent followed the waypoint forward. (rewarded 2.76)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
85% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, None, 'left')
Agent followed the waypoint left. (rewarded 2.61)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 1.53)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'red', 'forward', None, None)
Agent properly idled at a red light. (rewarded 2.42)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'red', 'left', None, 'left')
Agent attempted driving forward through a red light. (rewarded -10.42)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.
Agent previous state: ('forward', 'green', 'left', None, 'right')
Agent drove left instead of forward. (rewarded 1.02)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('right', 'red', None, None, 'right')
Agent followed the waypoint right. (rewarded 2.76)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('right', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 1.63)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('right', 'red', 'forward', None, 'right')
Agent properly idled at a red light. (rewarded 0.56)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('right', 'green', 'right', None, None)
Agent followed the waypoint right. (rewarded 1.13)
45% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 829
\-------------------------
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1909; alpha = 0.0020
/-------------------
| Step 0 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', 'forward')
Agent attempted driving left through a red light with traffic and cause a major accident. (rewarded -39.76)
95% of time remaining to reach destination.
/-------------------
| Step 1 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent properly idled at a red light. (rewarded 2.30)
90% of time remaining to reach destination.
/-------------------
| Step 2 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 2.01)
85% of time remaining to reach destination.
/-------------------
| Step 3 Results
\-------------------
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
Agent previous state: ('left', 'green', 'right', None, None)
Agent drove right instead of left. (rewarded 1.93)
80% of time remaining to reach destination.
/-------------------
| Step 4 Results
\-------------------
Agent previous state: ('forward', 'green', None, None, None)
Agent followed the waypoint forward. (rewarded 2.47)
75% of time remaining to reach destination.
/-------------------
| Step 5 Results
\-------------------
Agent previous state: ('forward', 'green', None, 'right', 'forward')
Agent followed the waypoint forward. (rewarded 1.30)
70% of time remaining to reach destination.
/-------------------
| Step 6 Results
\-------------------
Agent previous state: ('forward', 'red', None, 'forward', 'right')
Agent properly idled at a red light. (rewarded 2.87)
65% of time remaining to reach destination.
/-------------------
| Step 7 Results
\-------------------
Agent previous state: ('forward', 'red', 'right', 'left', None)
Agent properly idled at a red light. (rewarded 1.04)
60% of time remaining to reach destination.
/-------------------
| Step 8 Results
\-------------------
Agent previous state: ('forward', 'red', None, None, None)
Agent drove right instead of forward. (rewarded 0.07)
55% of time remaining to reach destination.
/-------------------
| Step 9 Results
\-------------------
Agent previous state: ('left', 'red', None, 'forward', None)
Agent properly idled at a red light. (rewarded 2.03)
50% of time remaining to reach destination.
/-------------------
| Step 10 Results
\-------------------
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
Agent previous state: ('left', 'green', None, 'right', None)
Agent followed the waypoint left. (rewarded 2.73)
45% of time remaining to reach destination.
/-------------------
| Step 11 Results
\-------------------
Agent previous state: ('left', 'red', None, None, 'left')
Agent properly idled at a red light. (rewarded 0.77)
40% of time remaining to reach destination.
/-------------------
| Step 12 Results
\-------------------
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.08)
35% of time remaining to reach destination.
Agent previous state: ('left', 'red', None, 'left', None)
Agent drove right instead of left. (rewarded 0.08)
35% of time remaining to reach destination.
/-------------------
| Step 13 Results
\-------------------
Agent previous state: ('right', 'red', 'left', None, None)
Agent followed the waypoint right. (rewarded 2.25)
30% of time remaining to reach destination.
/-------------------
| Step 14 Results
\-------------------
Agent previous state: ('right', 'red', None, 'right', None)
Agent followed the waypoint right. (rewarded 1.21)
25% of time remaining to reach destination.
/-------------------
| Step 15 Results
\-------------------
Agent previous state: ('right', 'red', 'right', None, None)
Agent followed the waypoint right. (rewarded 0.77)
20% of time remaining to reach destination.
/-------------------
| Step 16 Results
\-------------------
Agent previous state: ('left', 'red', None, None, None)
Agent properly idled at a red light. (rewarded 1.97)
15% of time remaining to reach destination.
/-------------------
| Step 17 Results
\-------------------
Agent previous state: ('left', 'green', None, None, None)
Agent followed the waypoint left. (rewarded 2.07)
10% of time remaining to reach destination.
Trial Completed!
Agent reached the destination.
/-------------------------
| Training trial 830
\-------------------------
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial. . .
epsilon = 0.1905; alpha = 0.0020
Simulating trial